code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_tf_bert import TFBertModelTester from ..clip.test_modeling_tf_clip import TFCLIPVisionModelTester from ..deit.test_modeling_tf_deit import TFDeiTModelTester from ..roberta.test_modeling_tf_roberta import TFRobertaModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester if is_tf_available(): from transformers import ( TFBertModel, TFCLIPVisionModel, TFDeiTModel, TFRobertaModel, TFVisionTextDualEncoderModel, TFViTModel, VisionTextDualEncoderConfig, ) if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor def __lowerCAmelCase (__lowerCAmelCase ): if isinstance(__lowerCAmelCase , collections.abc.Iterable ): return x return (x, x) @require_tf class lowerCAmelCase__ : def lowerCAmelCase__ ( self : str , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[Any] ) ->Optional[int]: '''simple docstring''' pass def lowerCAmelCase__ ( self : List[str] ) ->str: '''simple docstring''' pass def lowerCAmelCase__ ( self : Dict ) ->Dict: '''simple docstring''' pass def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : int=None , **lowerCamelCase__ : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = VisionTextDualEncoderConfig.from_vision_text_configs(__snake_case , __snake_case ) _UpperCAmelCase : Optional[int] = TFVisionTextDualEncoderModel(__snake_case ) _UpperCAmelCase : List[str] = model(input_ids=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim) ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : Any , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Dict=None , **lowerCamelCase__ : Any ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = self.get_vision_text_model(__snake_case , __snake_case ) _UpperCAmelCase : Optional[int] = TFVisionTextDualEncoderModel(vision_model=__snake_case , text_model=__snake_case ) _UpperCAmelCase : List[str] = model(input_ids=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : str , lowerCamelCase__ : Optional[Any]=None , **lowerCamelCase__ : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : str = self.get_vision_text_model(__snake_case , __snake_case ) _UpperCAmelCase : List[str] = {"vision_model": vision_model, "text_model": text_model} _UpperCAmelCase : Optional[Any] = TFVisionTextDualEncoderModel.from_vision_text_pretrained(**__snake_case ) _UpperCAmelCase : Union[str, Any] = model(input_ids=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : List[Any] , lowerCamelCase__ : int , lowerCamelCase__ : Union[str, Any]=None , **lowerCamelCase__ : Dict ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : int = self.get_vision_text_model(__snake_case , __snake_case ) _UpperCAmelCase : Tuple = TFVisionTextDualEncoderModel(vision_model=__snake_case , text_model=__snake_case ) _UpperCAmelCase : Dict = model(input_ids=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case ) _UpperCAmelCase : Tuple = output[0].numpy() with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__snake_case ) _UpperCAmelCase : Optional[Any] = TFVisionTextDualEncoderModel.from_pretrained(__snake_case ) _UpperCAmelCase : Tuple = model(input_ids=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case ) _UpperCAmelCase : Any = after_output[0].numpy() _UpperCAmelCase : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__snake_case , 1E-5 ) def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Tuple=None , **lowerCamelCase__ : Optional[int] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Dict = self.get_vision_text_model(__snake_case , __snake_case ) _UpperCAmelCase : Tuple = TFVisionTextDualEncoderModel(vision_model=__snake_case , text_model=__snake_case ) _UpperCAmelCase : Optional[Any] = model( input_ids=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case , output_attentions=__snake_case ) _UpperCAmelCase : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(__snake_case ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) _UpperCAmelCase : Optional[int] = to_atuple(vision_model.config.image_size ) _UpperCAmelCase : List[str] = to_atuple(vision_model.config.patch_size ) _UpperCAmelCase : int = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _UpperCAmelCase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _UpperCAmelCase : List[Any] = output.text_model_output.attentions self.assertEqual(len(__snake_case ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : float ) ->int: '''simple docstring''' _UpperCAmelCase : str = np.abs((a - b) ).max() self.assertLessEqual(__snake_case , __snake_case , F"""Difference between torch and flax is {diff} (>= {tol}).""" ) def lowerCAmelCase__ ( self : Optional[int] ) ->str: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_model(**__snake_case ) def lowerCAmelCase__ ( self : str ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Dict = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**__snake_case ) def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' _UpperCAmelCase : str = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**__snake_case ) def lowerCAmelCase__ ( self : int ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = self.prepare_config_and_inputs() self.check_save_load(**__snake_case ) def lowerCAmelCase__ ( self : List[str] ) ->int: '''simple docstring''' _UpperCAmelCase : Optional[int] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**__snake_case ) @slow def lowerCAmelCase__ ( self : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Dict = self.get_pretrained_model_and_inputs() _UpperCAmelCase : Optional[int] = model_a(**__snake_case ) _UpperCAmelCase : str = outputs[0].numpy() with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(__snake_case ) _UpperCAmelCase : Optional[Any] = TFVisionTextDualEncoderModel.from_pretrained(__snake_case ) _UpperCAmelCase : int = model_a(**__snake_case ) _UpperCAmelCase : Optional[Any] = after_outputs[0].numpy() _UpperCAmelCase : Optional[Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__snake_case , 1E-5 ) @require_tf class lowerCAmelCase__ ( lowerCamelCase_ , unittest.TestCase ): def lowerCAmelCase__ ( self : Tuple ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-random-bert" ) _UpperCAmelCase : int = 13 _UpperCAmelCase : str = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) _UpperCAmelCase : Optional[int] = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) _UpperCAmelCase : int = random_attention_mask([batch_size, 4] ) _UpperCAmelCase : Dict = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Tuple ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Tuple = TFViTModel(__snake_case , name="vision_model" ) _UpperCAmelCase : List[str] = TFBertModel(__snake_case , name="text_model" ) return vision_model, text_model def lowerCAmelCase__ ( self : str ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = TFViTModelTester(self ) _UpperCAmelCase : Optional[Any] = TFBertModelTester(self ) _UpperCAmelCase : int = vit_model_tester.prepare_config_and_inputs() _UpperCAmelCase : Tuple = bert_model_tester.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = vision_config_and_inputs ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) : List[str] = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class lowerCAmelCase__ ( lowerCamelCase_ , unittest.TestCase ): def lowerCAmelCase__ ( self : int ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : int = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "Rocketknight1/tiny-random-deit-tf" , "hf-internal-testing/tiny-random-roberta" ) _UpperCAmelCase : Dict = 13 _UpperCAmelCase : Tuple = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) _UpperCAmelCase : int = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) _UpperCAmelCase : Tuple = random_attention_mask([batch_size, 4] ) _UpperCAmelCase : int = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : int , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Dict=None , **lowerCamelCase__ : Optional[int] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : List[Any] = self.get_vision_text_model(__snake_case , __snake_case ) _UpperCAmelCase : Dict = TFVisionTextDualEncoderModel(vision_model=__snake_case , text_model=__snake_case ) _UpperCAmelCase : str = model( input_ids=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case , output_attentions=__snake_case ) _UpperCAmelCase : int = output.vision_model_output.attentions self.assertEqual(len(__snake_case ) , vision_config.num_hidden_layers ) # in DEiT, the seq_len equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _UpperCAmelCase : Union[str, Any] = to_atuple(vision_model.config.image_size ) _UpperCAmelCase : Any = to_atuple(vision_model.config.patch_size ) _UpperCAmelCase : Optional[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _UpperCAmelCase : List[Any] = num_patches + 2 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _UpperCAmelCase : Optional[Any] = output.text_model_output.attentions self.assertEqual(len(__snake_case ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : List[str] ) ->Any: '''simple docstring''' _UpperCAmelCase : Tuple = TFDeiTModel(__snake_case , name="vision_model" ) _UpperCAmelCase : Optional[Any] = TFRobertaModel(__snake_case , name="text_model" ) return vision_model, text_model def lowerCAmelCase__ ( self : Dict ) ->Dict: '''simple docstring''' _UpperCAmelCase : Dict = TFDeiTModelTester(self ) _UpperCAmelCase : List[str] = TFRobertaModelTester(self ) _UpperCAmelCase : Any = vit_model_tester.prepare_config_and_inputs() _UpperCAmelCase : Any = bert_model_tester.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = vision_config_and_inputs ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) : Any = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class lowerCAmelCase__ ( lowerCamelCase_ , unittest.TestCase ): def lowerCAmelCase__ ( self : List[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : int = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "Rocketknight1/tiny-random-clip-tf" , "hf-internal-testing/tiny-random-bert" ) _UpperCAmelCase : Optional[Any] = 13 _UpperCAmelCase : Dict = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) _UpperCAmelCase : List[str] = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) _UpperCAmelCase : str = random_attention_mask([batch_size, 4] ) _UpperCAmelCase : int = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : int = TFCLIPVisionModel(__snake_case , name="vision_model" ) _UpperCAmelCase : Optional[int] = TFBertModel(__snake_case , name="text_model" ) return vision_model, text_model def lowerCAmelCase__ ( self : Dict ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = TFCLIPVisionModelTester(self ) _UpperCAmelCase : Dict = TFBertModelTester(self ) _UpperCAmelCase : Union[str, Any] = clip_model_tester.prepare_config_and_inputs() _UpperCAmelCase : Optional[Any] = bert_model_tester.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase : List[str] = vision_config_and_inputs ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) : Dict = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_vision @require_tf class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Optional[Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[Any] = TFVisionTextDualEncoderModel.from_pretrained( "clip-italian/clip-italian" , logit_scale_init_value=1.0 , from_pt=__snake_case ) _UpperCAmelCase : int = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian" ) _UpperCAmelCase : Union[str, Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) _UpperCAmelCase : Union[str, Any] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=__snake_case , padding=__snake_case , return_tensors="np" ) _UpperCAmelCase : Dict = model(**__snake_case ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) _UpperCAmelCase : Tuple = np.array([[1.2_2_8_4_7_2_7, 0.3_1_0_4_1_2_2]] ) self.assertTrue(np.allclose(outputs.logits_per_image.numpy() , __snake_case , atol=1E-3 ) )
371
'''simple docstring''' import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCamelCase__ = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class lowerCAmelCase__ ( unittest.TestCase ): @classmethod def lowerCAmelCase__ ( cls : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : Tuple = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def lowerCAmelCase__ ( cls : Union[str, Any] ) ->int: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def lowerCAmelCase__ ( self : int ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ , repo_id="test-config" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : Dict = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ , repo_id="valid_org/test-config-org" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : int = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' CustomConfig.register_for_auto_class() _UpperCAmelCase : int = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) _UpperCAmelCase : str = AutoConfig.from_pretrained(F"""{USER}/test-dynamic-config""" , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : List[str] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated _UpperCAmelCase : Any = c.n_embd + 1 # int _UpperCAmelCase : List[Any] = c.resid_pdrop + 1.0 # float _UpperCAmelCase : Tuple = not c.scale_attn_weights # bool _UpperCAmelCase : List[Any] = c.summary_type + "foo" # str c.update_from_string( F"""n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}""" ) self.assertEqual(lowerCamelCase__ , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(lowerCamelCase__ , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(lowerCamelCase__ , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(lowerCamelCase__ , c.summary_type , "mismatch for key: summary_type" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = PretrainedConfig() _UpperCAmelCase : Tuple = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase__ , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) _UpperCAmelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase__ , lowerCamelCase__ )] if len(lowerCamelCase__ ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" F""" {', '.join(lowerCamelCase__ )}.""" ) def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): # config is in subfolder, the following should not work without specifying the subfolder _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = mock.Mock() _UpperCAmelCase : List[str] = 5_00 _UpperCAmelCase : Dict = {} _UpperCAmelCase : Tuple = HTTPError _UpperCAmelCase : Any = {} # Download this model to make sure it's in the cache. _UpperCAmelCase : int = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=lowerCamelCase__ ) as mock_head: _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = AutoConfig.from_pretrained("bert-base-cased" ) _UpperCAmelCase : str = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 _UpperCAmelCase : Optional[int] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 _UpperCAmelCase : Dict = ["config.42.0.0.json"] _UpperCAmelCase : Union[str, Any] = 7_68 configuration.save_pretrained(lowerCamelCase__ ) shutil.move(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , os.path.join(lowerCamelCase__ , "config.42.0.0.json" ) ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = "hf-internal-testing/test-two-configs" import transformers as new_transformers _UpperCAmelCase : Any = "v4.0.0" _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers _UpperCAmelCase : List[Any] = "v3.0.0" _UpperCAmelCase : int = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(old_configuration.hidden_size , 7_68 )
322
0
'''simple docstring''' from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class lowerCAmelCase__ ( UpperCAmelCase__ ): def __lt__( self : Any , lowerCamelCase__ : Tuple ) ->str: '''simple docstring''' return self[-1] < other[-1] def __eq__( self : Dict , lowerCamelCase__ : int ) ->Tuple: '''simple docstring''' return self[-1] == other[-1] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : list[Stack] = [] # sort into stacks for element in collection: _UpperCAmelCase : int = Stack([element] ) _UpperCAmelCase : Optional[int] = bisect_left(__lowerCAmelCase , __lowerCAmelCase ) if i != len(__lowerCAmelCase ): stacks[i].append(__lowerCAmelCase ) else: stacks.append(__lowerCAmelCase ) # use a heap-based merge to merge stack efficiently _UpperCAmelCase : List[str] = merge(*(reversed(__lowerCAmelCase ) for stack in stacks) ) return collection if __name__ == "__main__": lowerCamelCase__ = input('Enter numbers separated by a comma:\n').strip() lowerCamelCase__ = [int(item) for item in user_input.split(',')] print(patience_sort(unsorted))
350
'''simple docstring''' from manim import * class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : List[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Optional[Any] = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Dict = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("CPU" , font_size=24 ) _UpperCAmelCase : Any = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(1 )] _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("GPU" , font_size=24 ) _UpperCAmelCase : str = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) gpu.align_to(lowerCamelCase__ , lowerCamelCase__ ) gpu.set_x(gpu.get_x() - 1 ) self.add(lowerCamelCase__ ) _UpperCAmelCase : List[str] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Any = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Model" , font_size=24 ) _UpperCAmelCase : Tuple = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.play( Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , ) _UpperCAmelCase : int = MarkupText( F"""First, an empty model skeleton is loaded\ninto <span fgcolor='{YELLOW}'>memory</span> without using much RAM.""" , font_size=24 , ) _UpperCAmelCase : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ , run_time=2.5 ) , Write(lowerCamelCase__ ) , Write(lowerCamelCase__ ) ) self.add(lowerCamelCase__ ) _UpperCAmelCase : int = [] _UpperCAmelCase : List[str] = [] _UpperCAmelCase : Dict = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : int = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ , opacity=0.7 ) cpu_target.move_to(lowerCamelCase__ ) cpu_target.generate_target() _UpperCAmelCase : Dict = 0.4_6 / 4 _UpperCAmelCase : Any = 0.4_6 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowerCamelCase__ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=lowerCamelCase__ , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=lowerCamelCase__ , buff=0.0 ) cpu_targs.append(lowerCamelCase__ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(lowerCamelCase__ ) ) second_animations.append(MoveToTarget(lowerCamelCase__ , run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(*lowerCamelCase__ ) self.wait()
322
0
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase__ = logging.get_logger(__name__) # General docstring lowerCamelCase__ = 'RegNetConfig' # Base docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = [1, 1_088, 7, 7] # Image classification docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = 'tabby, tabby cat' lowerCamelCase__ = [ 'facebook/regnet-y-040', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 3 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[str] = "relu" , **lowerCamelCase__ : Tuple , ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _UpperCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _UpperCAmelCase : Dict = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=lowerCamelCase__ , strides=lowerCamelCase__ , padding="VALID" , groups=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" , ) _UpperCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) _UpperCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.convolution(self.padding(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = self.normalization(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = config.num_channels _UpperCAmelCase : Any = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = shape_list(lowerCamelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _UpperCAmelCase : Optional[Any] = tf.transpose(lowerCamelCase__ , perm=(0, 2, 3, 1) ) _UpperCAmelCase : List[Any] = self.embedder(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : int = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=1 , strides=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" ) _UpperCAmelCase : Any = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False ) ->tf.Tensor: '''simple docstring''' return self.normalization(self.convolution(lowerCamelCase__ ) , training=lowerCamelCase__ ) class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : int , **lowerCamelCase__ : Optional[int] ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) _UpperCAmelCase : int = [ tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.pooler(lowerCamelCase__ ) for layer_module in self.attention: _UpperCAmelCase : str = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : Any ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = in_channels != out_channels or stride != 1 _UpperCAmelCase : List[str] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : List[str] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _UpperCAmelCase : Optional[int] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.2" ), ] _UpperCAmelCase : Union[str, Any] = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = hidden_state for layer_module in self.layers: _UpperCAmelCase : List[Any] = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : List[Any] , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Optional[int] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : Union[str, Any] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) _UpperCAmelCase : List[Any] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(lowerCamelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.3" ), ] _UpperCAmelCase : int = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str ) ->Any: '''simple docstring''' _UpperCAmelCase : int = hidden_state for layer_module in self.layers: _UpperCAmelCase : Tuple = layer_module(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : Tuple = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : str = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer _UpperCAmelCase : List[str] = [ # downsampling is done in the first layer with stride of 2 layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , name="layers.0" ), *[layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , name=F"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] ) ->List[str]: '''simple docstring''' for layer_module in self.layers: _UpperCAmelCase : Optional[int] = layer_module(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : int ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) _UpperCAmelCase : Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCamelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , depth=lowerCamelCase__ , name=F"""stages.{i+1}""" ) ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = True ) ->TFBaseModelOutputWithNoAttention: '''simple docstring''' _UpperCAmelCase : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase : Optional[Any] = hidden_states + (hidden_state,) _UpperCAmelCase : Dict = stage_module(lowerCamelCase__ ) if output_hidden_states: _UpperCAmelCase : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase__ , hidden_states=lowerCamelCase__ ) @keras_serializable class lowerCAmelCase__ ( tf.keras.layers.Layer ): lowerCAmelCase : Optional[Any] = RegNetConfig def __init__( self : Union[str, Any] , lowerCamelCase__ : Any , **lowerCamelCase__ : str ) ->int: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = config _UpperCAmelCase : Union[str, Any] = TFRegNetEmbeddings(lowerCamelCase__ , name="embedder" ) _UpperCAmelCase : Union[str, Any] = TFRegNetEncoder(lowerCamelCase__ , name="encoder" ) _UpperCAmelCase : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) @unpack_inputs def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : bool = False , ) ->TFBaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' _UpperCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.embedder(lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : str = self.encoder( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : Dict = encoder_outputs[0] _UpperCAmelCase : Dict = self.pooler(lowerCamelCase__ ) # Change to NCHW output format have uniformity in the modules _UpperCAmelCase : Union[str, Any] = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) _UpperCAmelCase : Tuple = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _UpperCAmelCase : List[str] = tuple([tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase__ , pooler_output=lowerCamelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = RegNetConfig lowerCAmelCase : Tuple = "regnet" lowerCAmelCase : Union[str, Any] = "pixel_values" @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} lowerCamelCase__ = r'\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCamelCase__ = r'\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : Any , **lowerCamelCase__ : List[str] ) ->Optional[int]: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Any=False , ) ->Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( pixel_values=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : List[Any] , **lowerCamelCase__ : Union[str, Any] ) ->Any: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = config.num_labels _UpperCAmelCase : Dict = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) # classification head _UpperCAmelCase : str = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict=False , ) ->Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : str = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : int = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase : Dict = self.classifier[0](lowerCamelCase__ ) _UpperCAmelCase : str = self.classifier[1](lowerCamelCase__ ) _UpperCAmelCase : Tuple = None if labels is None else self.hf_compute_loss(labels=lowerCamelCase__ , logits=lowerCamelCase__ ) if not return_dict: _UpperCAmelCase : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCamelCase__ , logits=lowerCamelCase__ , hidden_states=outputs.hidden_states )
351
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1_024 , __lowerCAmelCase=1_024 , __lowerCAmelCase=False , **__lowerCAmelCase ): _UpperCAmelCase : Any = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : List[str] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="train" , **__lowerCAmelCase ) _UpperCAmelCase : Dict = tok.pad_token_id def get_lens(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = tqdm( DataLoader(__lowerCAmelCase , batch_size=512 , num_workers=8 , shuffle=__lowerCAmelCase , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) _UpperCAmelCase : List[str] = [] for batch in dl: _UpperCAmelCase : Any = batch["input_ids"].ne(__lowerCAmelCase ).sum(1 ).tolist() _UpperCAmelCase : Tuple = batch["labels"].ne(__lowerCAmelCase ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__lowerCAmelCase , __lowerCAmelCase ): max_lens.append(max(__lowerCAmelCase , __lowerCAmelCase ) ) else: max_lens.extend(__lowerCAmelCase ) return max_lens _UpperCAmelCase : Dict = get_lens(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="val" , **__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = get_lens(__lowerCAmelCase ) pickle_save(__lowerCAmelCase , train_ds.len_file ) pickle_save(__lowerCAmelCase , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
322
0
'''simple docstring''' from manim import * class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Any ) ->Any: '''simple docstring''' _UpperCAmelCase : Optional[int] = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Dict = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _UpperCAmelCase : Dict = [mem.copy() for i in range(6 )] _UpperCAmelCase : int = [mem.copy() for i in range(6 )] _UpperCAmelCase : List[Any] = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : List[str] = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = Text("CPU" , font_size=24 ) _UpperCAmelCase : Optional[int] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : str = [mem.copy() for i in range(4 )] _UpperCAmelCase : Optional[int] = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Dict = Text("GPU" , font_size=24 ) _UpperCAmelCase : List[Any] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) gpu.move_to([-1, -1, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Tuple = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Model" , font_size=24 ) _UpperCAmelCase : int = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = [] for i, rect in enumerate(lowerCamelCase__ ): rect.set_stroke(lowerCamelCase__ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _UpperCAmelCase : Optional[int] = Rectangle(height=0.4_6 / 4 , width=0.4_6 / 3 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowerCamelCase__ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowerCamelCase__ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowerCamelCase__ , buff=0.0 ) self.add(lowerCamelCase__ ) cpu_targs.append(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Tuple = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Loaded Checkpoint" , font_size=24 ) _UpperCAmelCase : Optional[int] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , aligned_edge=lowerCamelCase__ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _UpperCAmelCase : Tuple = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCAmelCase : Optional[Any] = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowerCamelCase__ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ ) , Write(lowerCamelCase__ ) ) self.play(Write(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) ) _UpperCAmelCase : Any = [] _UpperCAmelCase : str = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : Dict = fill.copy().set_fill(lowerCamelCase__ , opacity=0.7 ) target.move_to(lowerCamelCase__ ) first_animations.append(GrowFromCenter(lowerCamelCase__ , run_time=1 ) ) _UpperCAmelCase : Dict = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(lowerCamelCase__ , run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(*lowerCamelCase__ ) self.wait()
352
'''simple docstring''' import pytest lowerCamelCase__ = '__dummy_dataset1__' lowerCamelCase__ = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = dataset_loading_script_name _UpperCAmelCase : Any = tmp_path / "datasets" / script_name script_dir.mkdir(parents=__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = script_dir / F"""{script_name}.py""" with open(__lowerCAmelCase , "w" ) as f: f.write(__lowerCAmelCase ) return str(__lowerCAmelCase )
322
0
'''simple docstring''' import gc import threading import time import psutil import torch class lowerCAmelCase__ : def __init__( self : str ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : int = psutil.Process() _UpperCAmelCase : int = False def lowerCAmelCase__ ( self : Union[str, Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : Tuple = -1 while True: _UpperCAmelCase : Optional[int] = max(self.process.memory_info().rss , self.cpu_memory_peak ) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def lowerCAmelCase__ ( self : List[Any] ) ->int: '''simple docstring''' _UpperCAmelCase : Any = True _UpperCAmelCase : Union[str, Any] = threading.Thread(target=self.peak_monitor ) _UpperCAmelCase : Any = True self.thread.start() def lowerCAmelCase__ ( self : Union[str, Any] ) ->int: '''simple docstring''' _UpperCAmelCase : Any = False self.thread.join() return self.cpu_memory_peak lowerCamelCase__ = PeakCPUMemory() def __lowerCAmelCase () -> Dict: # Time _UpperCAmelCase : Dict = {"time": time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem _UpperCAmelCase : List[str] = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): _UpperCAmelCase : Union[str, Any] = torch.cuda.memory_allocated(__lowerCAmelCase ) torch.cuda.reset_peak_memory_stats() return measures def __lowerCAmelCase (__lowerCAmelCase ) -> str: # Time _UpperCAmelCase : Union[str, Any] = {"time": time.time() - start_measures["time"]} gc.collect() torch.cuda.empty_cache() # CPU mem _UpperCAmelCase : List[Any] = (psutil.Process().memory_info().rss - start_measures["cpu"]) / 2**20 _UpperCAmelCase : Union[str, Any] = (cpu_peak_tracker.stop() - start_measures["cpu"]) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): _UpperCAmelCase : List[str] = (torch.cuda.memory_allocated(__lowerCAmelCase ) - start_measures[str(__lowerCAmelCase )]) / 2**20 _UpperCAmelCase : Tuple = (torch.cuda.max_memory_allocated(__lowerCAmelCase ) - start_measures[str(__lowerCAmelCase )]) / 2**20 return measures def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ) -> str: print(F"""{description}:""" ) print(F"""- Time: {measures['time']:.2f}s""" ) for i in range(torch.cuda.device_count() ): print(F"""- GPU {i} allocated: {measures[str(__lowerCAmelCase )]:.2f}MiB""" ) _UpperCAmelCase : Union[str, Any] = measures[F"""{i}-peak"""] print(F"""- GPU {i} peak: {peak:.2f}MiB""" ) print(F"""- CPU RAM allocated: {measures['cpu']:.2f}MiB""" ) print(F"""- CPU RAM peak: {measures['cpu-peak']:.2f}MiB""" )
353
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase__ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCamelCase__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCamelCase__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCamelCase__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"] , reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ] , ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] ) ->int: '''simple docstring''' import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : int=0.9 , lowerCamelCase__ : Dict=3 , lowerCamelCase__ : Dict=0.5 ) ->Any: '''simple docstring''' if NLTK_VERSION >= version.Version("3.6.5" ): _UpperCAmelCase : Dict = [ meteor_score.single_meteor_score( word_tokenize(lowerCamelCase__ ) , word_tokenize(lowerCamelCase__ ) , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] else: _UpperCAmelCase : Optional[int] = [ meteor_score.single_meteor_score(lowerCamelCase__ , lowerCamelCase__ , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] return {"meteor": np.mean(lowerCamelCase__ )}
322
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'kssteven/ibert-roberta-base': 'https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json', 'kssteven/ibert-roberta-large': 'https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json', 'kssteven/ibert-roberta-large-mnli': ( 'https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json' ), } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = "ibert" def __init__( self : List[Any] , lowerCamelCase__ : Dict=3_05_22 , lowerCamelCase__ : Dict=7_68 , lowerCamelCase__ : Union[str, Any]=12 , lowerCamelCase__ : Tuple=12 , lowerCamelCase__ : List[Any]=30_72 , lowerCamelCase__ : Any="gelu" , lowerCamelCase__ : int=0.1 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : str=5_12 , lowerCamelCase__ : Dict=2 , lowerCamelCase__ : List[str]=0.0_2 , lowerCamelCase__ : Optional[Any]=1E-12 , lowerCamelCase__ : List[str]=1 , lowerCamelCase__ : Tuple=0 , lowerCamelCase__ : Tuple=2 , lowerCamelCase__ : Tuple="absolute" , lowerCamelCase__ : Optional[Any]=False , lowerCamelCase__ : Dict="none" , **lowerCamelCase__ : str , ) ->int: '''simple docstring''' super().__init__(pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Tuple = vocab_size _UpperCAmelCase : Dict = hidden_size _UpperCAmelCase : int = num_hidden_layers _UpperCAmelCase : str = num_attention_heads _UpperCAmelCase : Optional[int] = hidden_act _UpperCAmelCase : List[str] = intermediate_size _UpperCAmelCase : List[Any] = hidden_dropout_prob _UpperCAmelCase : Optional[Any] = attention_probs_dropout_prob _UpperCAmelCase : List[Any] = max_position_embeddings _UpperCAmelCase : Dict = type_vocab_size _UpperCAmelCase : List[Any] = initializer_range _UpperCAmelCase : Optional[Any] = layer_norm_eps _UpperCAmelCase : Optional[int] = position_embedding_type _UpperCAmelCase : Dict = quant_mode _UpperCAmelCase : str = force_dequant class lowerCAmelCase__ ( UpperCAmelCase__ ): @property def lowerCAmelCase__ ( self : Dict ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": _UpperCAmelCase : List[Any] = {0: "batch", 1: "choice", 2: "sequence"} else: _UpperCAmelCase : List[str] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
354
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline lowerCamelCase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : int ) ->str: '''simple docstring''' if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Union[str, Any] = [label.strip() for label in labels.split("," ) if label.strip()] return labels def __call__( self : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] ) ->str: '''simple docstring''' if len(lowerCamelCase__ ) == 0 or len(lowerCamelCase__ ) == 0: raise ValueError("You must include at least one label and at least one sequence." ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( "The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. " "Make sure the passed template includes formatting syntax such as {{}} where the label should go." ).format(lowerCamelCase__ ) ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Optional[Any] = [sequences] _UpperCAmelCase : int = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(lowerCamelCase__ )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(UpperCAmelCase__ ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Union[str, Any] , lowerCamelCase__ : Optional[Any]=ZeroShotClassificationArgumentHandler() , *lowerCamelCase__ : List[str] , **lowerCamelCase__ : Any ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = args_parser super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) if self.entailment_id == -1: logger.warning( "Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to " "-1. Define a descriptive label2id mapping in the model config to ensure correct outputs." ) @property def lowerCAmelCase__ ( self : Any ) ->Union[str, Any]: '''simple docstring''' for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("entail" ): return ind return -1 def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Tuple , lowerCamelCase__ : int=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : str=TruncationStrategy.ONLY_FIRST , **lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : int = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( "Tokenizer was not supporting padding necessary for zero-shot, attempting to use " " `pad_token=eos_token`" ) _UpperCAmelCase : Optional[Any] = self.tokenizer.eos_token try: _UpperCAmelCase : List[str] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , ) except Exception as e: if "too short" in str(lowerCamelCase__ ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. _UpperCAmelCase : List[Any] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def lowerCAmelCase__ ( self : int , **lowerCamelCase__ : Union[str, Any] ) ->Tuple: '''simple docstring''' if kwargs.get("multi_class" , lowerCamelCase__ ) is not None: _UpperCAmelCase : int = kwargs["multi_class"] logger.warning( "The `multi_class` argument has been deprecated and renamed to `multi_label`. " "`multi_class` will be removed in a future version of Transformers." ) _UpperCAmelCase : Dict = {} if "candidate_labels" in kwargs: _UpperCAmelCase : List[Any] = self._args_parser._parse_labels(kwargs["candidate_labels"] ) if "hypothesis_template" in kwargs: _UpperCAmelCase : Dict = kwargs["hypothesis_template"] _UpperCAmelCase : List[str] = {} if "multi_label" in kwargs: _UpperCAmelCase : Optional[Any] = kwargs["multi_label"] return preprocess_params, {}, postprocess_params def __call__( self : int , lowerCamelCase__ : Union[str, List[str]] , *lowerCamelCase__ : str , **lowerCamelCase__ : Optional[Any] , ) ->Optional[int]: '''simple docstring''' if len(lowerCamelCase__ ) == 0: pass elif len(lowerCamelCase__ ) == 1 and "candidate_labels" not in kwargs: _UpperCAmelCase : int = args[0] else: raise ValueError(F"""Unable to understand extra arguments {args}""" ) return super().__call__(lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Any=None , lowerCamelCase__ : str="This example is {}." ) ->Tuple: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Optional[int] = self._args_parser(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for i, (candidate_label, sequence_pair) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): _UpperCAmelCase : Optional[int] = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(lowerCamelCase__ ) - 1, **model_input, } def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Optional[int] ) ->int: '''simple docstring''' _UpperCAmelCase : Dict = inputs["candidate_label"] _UpperCAmelCase : Optional[int] = inputs["sequence"] _UpperCAmelCase : Dict = {k: inputs[k] for k in self.tokenizer.model_input_names} _UpperCAmelCase : List[Any] = self.model(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = { "candidate_label": candidate_label, "sequence": sequence, "is_last": inputs["is_last"], **outputs, } return model_outputs def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple=False ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = [outputs["candidate_label"] for outputs in model_outputs] _UpperCAmelCase : Any = [outputs["sequence"] for outputs in model_outputs] _UpperCAmelCase : Optional[int] = np.concatenate([output["logits"].numpy() for output in model_outputs] ) _UpperCAmelCase : Optional[Any] = logits.shape[0] _UpperCAmelCase : Any = len(lowerCamelCase__ ) _UpperCAmelCase : str = N // n _UpperCAmelCase : str = logits.reshape((num_sequences, n, -1) ) if multi_label or len(lowerCamelCase__ ) == 1: # softmax over the entailment vs. contradiction dim for each label independently _UpperCAmelCase : int = self.entailment_id _UpperCAmelCase : List[Any] = -1 if entailment_id == 0 else 0 _UpperCAmelCase : str = reshaped_outputs[..., [contradiction_id, entailment_id]] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : str = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels _UpperCAmelCase : int = reshaped_outputs[..., self.entailment_id] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
322
0
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase__ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCamelCase__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCamelCase__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCamelCase__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"] , reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ] , ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] ) ->int: '''simple docstring''' import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : int=0.9 , lowerCamelCase__ : Dict=3 , lowerCamelCase__ : Dict=0.5 ) ->Any: '''simple docstring''' if NLTK_VERSION >= version.Version("3.6.5" ): _UpperCAmelCase : Dict = [ meteor_score.single_meteor_score( word_tokenize(lowerCamelCase__ ) , word_tokenize(lowerCamelCase__ ) , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] else: _UpperCAmelCase : Optional[int] = [ meteor_score.single_meteor_score(lowerCamelCase__ , lowerCamelCase__ , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] return {"meteor": np.mean(lowerCamelCase__ )}
355
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase = 4_000_000 ): _UpperCAmelCase : List[Any] = [] _UpperCAmelCase , _UpperCAmelCase : Dict = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__lowerCAmelCase ) _UpperCAmelCase , _UpperCAmelCase : Any = b, a + b return sum(__lowerCAmelCase ) if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def __lowerCAmelCase (__lowerCAmelCase ): if is_torch_version("<" , "2.0.0" ) or not hasattr(__lowerCAmelCase , "_dynamo" ): return False return isinstance(__lowerCAmelCase , torch._dynamo.eval_frame.OptimizedModule ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase = True ): _UpperCAmelCase : Any = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) _UpperCAmelCase : Dict = is_compiled_module(__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : Optional[int] = model _UpperCAmelCase : Any = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = model.module if not keep_fpaa_wrapper: _UpperCAmelCase : List[Any] = getattr(__lowerCAmelCase , "forward" ) _UpperCAmelCase : Dict = model.__dict__.pop("_original_forward" , __lowerCAmelCase ) if original_forward is not None: while hasattr(__lowerCAmelCase , "__wrapped__" ): _UpperCAmelCase : Optional[int] = forward.__wrapped__ if forward == original_forward: break _UpperCAmelCase : Dict = forward if getattr(__lowerCAmelCase , "_converted_to_transformer_engine" , __lowerCAmelCase ): convert_model(__lowerCAmelCase , to_transformer_engine=__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : int = model _UpperCAmelCase : str = compiled_model return model def __lowerCAmelCase (): PartialState().wait_for_everyone() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(__lowerCAmelCase , __lowerCAmelCase ) elif PartialState().local_process_index == 0: torch.save(__lowerCAmelCase , __lowerCAmelCase ) @contextmanager def __lowerCAmelCase (**__lowerCAmelCase ): for key, value in kwargs.items(): _UpperCAmelCase : str = str(__lowerCAmelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def __lowerCAmelCase (__lowerCAmelCase ): if not hasattr(__lowerCAmelCase , "__qualname__" ) and not hasattr(__lowerCAmelCase , "__name__" ): _UpperCAmelCase : List[str] = getattr(__lowerCAmelCase , "__class__" , __lowerCAmelCase ) if hasattr(__lowerCAmelCase , "__qualname__" ): return obj.__qualname__ if hasattr(__lowerCAmelCase , "__name__" ): return obj.__name__ return str(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): for key, value in source.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = destination.setdefault(__lowerCAmelCase , {} ) merge_dicts(__lowerCAmelCase , __lowerCAmelCase ) else: _UpperCAmelCase : Optional[int] = value return destination def __lowerCAmelCase (__lowerCAmelCase = None ): if port is None: _UpperCAmelCase : Tuple = 29_500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
356
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[str]=13 , lowerCamelCase__ : Optional[Any]=7 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : int=99 , lowerCamelCase__ : int=32 , lowerCamelCase__ : List[str]=5 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : Optional[int]=37 , lowerCamelCase__ : Tuple="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Union[str, Any]=0.1 , lowerCamelCase__ : Optional[int]=5_12 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : str=2 , lowerCamelCase__ : Union[str, Any]=0.0_2 , lowerCamelCase__ : Tuple=4 , ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = parent _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Optional[int] = seq_length _UpperCAmelCase : int = is_training _UpperCAmelCase : Dict = use_attention_mask _UpperCAmelCase : Optional[Any] = use_token_type_ids _UpperCAmelCase : int = use_labels _UpperCAmelCase : Optional[int] = vocab_size _UpperCAmelCase : Any = hidden_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : List[Any] = num_attention_heads _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : int = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob _UpperCAmelCase : Union[str, Any] = max_position_embeddings _UpperCAmelCase : Tuple = type_vocab_size _UpperCAmelCase : List[Any] = type_sequence_label_size _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Dict = num_choices def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Dict = None if self.use_attention_mask: _UpperCAmelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCAmelCase : Union[str, Any] = None if self.use_token_type_ids: _UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCAmelCase : int = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = config_and_inputs _UpperCAmelCase : str = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Optional[int] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase__ ( self : Optional[int] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = FlaxAlbertModelTester(self ) @slow def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCAmelCase : List[str] = model_class_name.from_pretrained("albert-base-v2" ) _UpperCAmelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = FlaxAlbertModel.from_pretrained("albert-base-v2" ) _UpperCAmelCase : List[Any] = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) _UpperCAmelCase : Optional[int] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCAmelCase : Dict = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )[0] _UpperCAmelCase : List[Any] = (1, 11, 7_68) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : str = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase__ , atol=1E-4 ) )
322
0
'''simple docstring''' from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image lowerCamelCase__ : int = ['text', 'image', 'audio'] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Optional[int] = [] for input_type in input_types: if input_type == "text": inputs.append("Text input" ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" ).resize((512, 512) ) ) elif input_type == "audio": inputs.append(torch.ones(3_000 ) ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): inputs.append(create_inputs(__lowerCAmelCase ) ) else: raise ValueError(F"""Invalid type requested: {input_type}""" ) return inputs def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Any = [] for output in outputs: if isinstance(__lowerCAmelCase , (str, AgentText) ): output_types.append("text" ) elif isinstance(__lowerCAmelCase , (Image.Image, AgentImage) ): output_types.append("image" ) elif isinstance(__lowerCAmelCase , (torch.Tensor, AgentAudio) ): output_types.append("audio" ) else: raise ValueError(F"""Invalid output: {output}""" ) return output_types @is_tool_test class lowerCAmelCase__ : def lowerCAmelCase__ ( self : Optional[Any] ) ->Tuple: '''simple docstring''' self.assertTrue(hasattr(self.tool , "inputs" ) ) self.assertTrue(hasattr(self.tool , "outputs" ) ) _UpperCAmelCase : int = self.tool.inputs for _input in inputs: if isinstance(_input , lowerCamelCase__ ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) _UpperCAmelCase : int = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def lowerCAmelCase__ ( self : Union[str, Any] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = create_inputs(self.tool.inputs ) _UpperCAmelCase : Tuple = self.tool(*lowerCamelCase__ ) # There is a single output if len(self.tool.outputs ) == 1: _UpperCAmelCase : Union[str, Any] = [outputs] self.assertListEqual(output_types(lowerCamelCase__ ) , self.tool.outputs ) def lowerCAmelCase__ ( self : Dict ) ->Tuple: '''simple docstring''' self.assertTrue(hasattr(self.tool , "description" ) ) self.assertTrue(hasattr(self.tool , "default_checkpoint" ) ) self.assertTrue(self.tool.description.startswith("This is a tool that" ) ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[Any] = create_inputs(self.tool.inputs ) _UpperCAmelCase : str = self.tool(*lowerCamelCase__ ) if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : int = [outputs] self.assertEqual(len(lowerCamelCase__ ) , len(self.tool.outputs ) ) for output, output_type in zip(lowerCamelCase__ , self.tool.outputs ): _UpperCAmelCase : List[Any] = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[Any] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = create_inputs(self.tool.inputs ) _UpperCAmelCase : List[Any] = [] for _input, input_type in zip(lowerCamelCase__ , self.tool.inputs ): if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error _UpperCAmelCase : List[Any] = self.tool(*lowerCamelCase__ ) if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : List[Any] = [outputs] self.assertEqual(len(lowerCamelCase__ ) , len(self.tool.outputs ) )
357
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-classification/requirements.txt') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "rb" ) as f: _UpperCAmelCase : List[str] = Image.open(__lowerCAmelCase ) return im.convert("RGB" ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={ "help": "Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub)." } , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the training data."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the validation data."} ) lowerCAmelCase : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default="google/vit-base-patch16-224-in21k" , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCAmelCase__ )} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) lowerCAmelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCAmelCase : str = field(default=UpperCAmelCase__ , metadata={"help": "Name or path of preprocessor config."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = torch.stack([example["pixel_values"] for example in examples] ) _UpperCAmelCase : Tuple = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def __lowerCAmelCase (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , __lowerCAmelCase , __lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(__lowerCAmelCase ) transformers.utils.logging.set_verbosity(__lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _UpperCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCAmelCase : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCAmelCase : List[Any] = {} if data_args.train_dir is not None: _UpperCAmelCase : str = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCAmelCase : Optional[Any] = os.path.join(data_args.validation_dir , "**" ) _UpperCAmelCase : Any = load_dataset( "imagefolder" , data_files=__lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCAmelCase : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __lowerCAmelCase ) and data_args.train_val_split > 0.0: _UpperCAmelCase : List[Any] = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCAmelCase : List[str] = split["train"] _UpperCAmelCase : Union[str, Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase : Optional[int] = dataset["train"].features["labels"].names _UpperCAmelCase , _UpperCAmelCase : int = {}, {} for i, label in enumerate(__lowerCAmelCase ): _UpperCAmelCase : int = str(__lowerCAmelCase ) _UpperCAmelCase : str = label # Load the accuracy metric from the datasets package _UpperCAmelCase : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(__lowerCAmelCase ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowerCAmelCase ) , labelaid=__lowerCAmelCase , idalabel=__lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCAmelCase : Dict = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCAmelCase : int = image_processor.size["shortest_edge"] else: _UpperCAmelCase : int = (image_processor.size["height"], image_processor.size["width"]) _UpperCAmelCase : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCAmelCase : Optional[int] = Compose( [ RandomResizedCrop(__lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCAmelCase : Union[str, Any] = Compose( [ Resize(__lowerCAmelCase ), CenterCrop(__lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCAmelCase : Dict = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(__lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCAmelCase : Optional[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(__lowerCAmelCase ) # Initalize our trainer _UpperCAmelCase : Union[str, Any] = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=__lowerCAmelCase , tokenizer=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : Any = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : int = last_checkpoint _UpperCAmelCase : Dict = trainer.train(resume_from_checkpoint=__lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowerCAmelCase ) trainer.save_metrics("eval" , __lowerCAmelCase ) # Write model card and (optionally) push to hub _UpperCAmelCase : int = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowerCAmelCase ) else: trainer.create_model_card(**__lowerCAmelCase ) if __name__ == "__main__": main()
322
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { 'configuration_blip_2': [ 'BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Blip2Config', 'Blip2QFormerConfig', 'Blip2VisionConfig', ], 'processing_blip_2': ['Blip2Processor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ 'BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST', 'Blip2Model', 'Blip2QFormerModel', 'Blip2PreTrainedModel', 'Blip2ForConditionalGeneration', 'Blip2VisionModel', ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
358
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase__ = logging.get_logger(__name__) # General docstring lowerCamelCase__ = 'RegNetConfig' # Base docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = [1, 1_088, 7, 7] # Image classification docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = 'tabby, tabby cat' lowerCamelCase__ = [ 'facebook/regnet-y-040', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 3 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[str] = "relu" , **lowerCamelCase__ : Tuple , ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _UpperCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _UpperCAmelCase : Dict = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=lowerCamelCase__ , strides=lowerCamelCase__ , padding="VALID" , groups=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" , ) _UpperCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) _UpperCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.convolution(self.padding(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = self.normalization(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = config.num_channels _UpperCAmelCase : Any = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = shape_list(lowerCamelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _UpperCAmelCase : Optional[Any] = tf.transpose(lowerCamelCase__ , perm=(0, 2, 3, 1) ) _UpperCAmelCase : List[Any] = self.embedder(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : int = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=1 , strides=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" ) _UpperCAmelCase : Any = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False ) ->tf.Tensor: '''simple docstring''' return self.normalization(self.convolution(lowerCamelCase__ ) , training=lowerCamelCase__ ) class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : int , **lowerCamelCase__ : Optional[int] ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) _UpperCAmelCase : int = [ tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.pooler(lowerCamelCase__ ) for layer_module in self.attention: _UpperCAmelCase : str = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : Any ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = in_channels != out_channels or stride != 1 _UpperCAmelCase : List[str] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : List[str] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _UpperCAmelCase : Optional[int] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.2" ), ] _UpperCAmelCase : Union[str, Any] = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = hidden_state for layer_module in self.layers: _UpperCAmelCase : List[Any] = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : List[Any] , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Optional[int] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : Union[str, Any] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) _UpperCAmelCase : List[Any] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(lowerCamelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.3" ), ] _UpperCAmelCase : int = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str ) ->Any: '''simple docstring''' _UpperCAmelCase : int = hidden_state for layer_module in self.layers: _UpperCAmelCase : Tuple = layer_module(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : Tuple = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : str = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer _UpperCAmelCase : List[str] = [ # downsampling is done in the first layer with stride of 2 layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , name="layers.0" ), *[layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , name=F"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] ) ->List[str]: '''simple docstring''' for layer_module in self.layers: _UpperCAmelCase : Optional[int] = layer_module(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : int ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) _UpperCAmelCase : Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCamelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , depth=lowerCamelCase__ , name=F"""stages.{i+1}""" ) ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = True ) ->TFBaseModelOutputWithNoAttention: '''simple docstring''' _UpperCAmelCase : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase : Optional[Any] = hidden_states + (hidden_state,) _UpperCAmelCase : Dict = stage_module(lowerCamelCase__ ) if output_hidden_states: _UpperCAmelCase : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase__ , hidden_states=lowerCamelCase__ ) @keras_serializable class lowerCAmelCase__ ( tf.keras.layers.Layer ): lowerCAmelCase : Optional[Any] = RegNetConfig def __init__( self : Union[str, Any] , lowerCamelCase__ : Any , **lowerCamelCase__ : str ) ->int: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = config _UpperCAmelCase : Union[str, Any] = TFRegNetEmbeddings(lowerCamelCase__ , name="embedder" ) _UpperCAmelCase : Union[str, Any] = TFRegNetEncoder(lowerCamelCase__ , name="encoder" ) _UpperCAmelCase : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) @unpack_inputs def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : bool = False , ) ->TFBaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' _UpperCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.embedder(lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : str = self.encoder( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : Dict = encoder_outputs[0] _UpperCAmelCase : Dict = self.pooler(lowerCamelCase__ ) # Change to NCHW output format have uniformity in the modules _UpperCAmelCase : Union[str, Any] = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) _UpperCAmelCase : Tuple = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _UpperCAmelCase : List[str] = tuple([tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase__ , pooler_output=lowerCamelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = RegNetConfig lowerCAmelCase : Tuple = "regnet" lowerCAmelCase : Union[str, Any] = "pixel_values" @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} lowerCamelCase__ = r'\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCamelCase__ = r'\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : Any , **lowerCamelCase__ : List[str] ) ->Optional[int]: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Any=False , ) ->Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( pixel_values=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : List[Any] , **lowerCamelCase__ : Union[str, Any] ) ->Any: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = config.num_labels _UpperCAmelCase : Dict = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) # classification head _UpperCAmelCase : str = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict=False , ) ->Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : str = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : int = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase : Dict = self.classifier[0](lowerCamelCase__ ) _UpperCAmelCase : str = self.classifier[1](lowerCamelCase__ ) _UpperCAmelCase : Tuple = None if labels is None else self.hf_compute_loss(labels=lowerCamelCase__ , logits=lowerCamelCase__ ) if not return_dict: _UpperCAmelCase : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCamelCase__ , logits=lowerCamelCase__ , hidden_states=outputs.hidden_states )
322
0
'''simple docstring''' import math from numpy import inf from scipy.integrate import quad def __lowerCAmelCase (__lowerCAmelCase ): if num <= 0: raise ValueError("math domain error" ) return quad(__lowerCAmelCase , 0 , __lowerCAmelCase , args=(__lowerCAmelCase) )[0] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return math.pow(__lowerCAmelCase , z - 1 ) * math.exp(-x ) if __name__ == "__main__": from doctest import testmod testmod()
359
'''simple docstring''' import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def __lowerCAmelCase (__lowerCAmelCase ): if is_torch_version("<" , "2.0.0" ) or not hasattr(__lowerCAmelCase , "_dynamo" ): return False return isinstance(__lowerCAmelCase , torch._dynamo.eval_frame.OptimizedModule ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase = True ): _UpperCAmelCase : Any = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) _UpperCAmelCase : Dict = is_compiled_module(__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : Optional[int] = model _UpperCAmelCase : Any = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = model.module if not keep_fpaa_wrapper: _UpperCAmelCase : List[Any] = getattr(__lowerCAmelCase , "forward" ) _UpperCAmelCase : Dict = model.__dict__.pop("_original_forward" , __lowerCAmelCase ) if original_forward is not None: while hasattr(__lowerCAmelCase , "__wrapped__" ): _UpperCAmelCase : Optional[int] = forward.__wrapped__ if forward == original_forward: break _UpperCAmelCase : Dict = forward if getattr(__lowerCAmelCase , "_converted_to_transformer_engine" , __lowerCAmelCase ): convert_model(__lowerCAmelCase , to_transformer_engine=__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : int = model _UpperCAmelCase : str = compiled_model return model def __lowerCAmelCase (): PartialState().wait_for_everyone() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(__lowerCAmelCase , __lowerCAmelCase ) elif PartialState().local_process_index == 0: torch.save(__lowerCAmelCase , __lowerCAmelCase ) @contextmanager def __lowerCAmelCase (**__lowerCAmelCase ): for key, value in kwargs.items(): _UpperCAmelCase : str = str(__lowerCAmelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def __lowerCAmelCase (__lowerCAmelCase ): if not hasattr(__lowerCAmelCase , "__qualname__" ) and not hasattr(__lowerCAmelCase , "__name__" ): _UpperCAmelCase : List[str] = getattr(__lowerCAmelCase , "__class__" , __lowerCAmelCase ) if hasattr(__lowerCAmelCase , "__qualname__" ): return obj.__qualname__ if hasattr(__lowerCAmelCase , "__name__" ): return obj.__name__ return str(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): for key, value in source.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = destination.setdefault(__lowerCAmelCase , {} ) merge_dicts(__lowerCAmelCase , __lowerCAmelCase ) else: _UpperCAmelCase : Optional[int] = value return destination def __lowerCAmelCase (__lowerCAmelCase = None ): if port is None: _UpperCAmelCase : Tuple = 29_500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
322
0
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[Any] = filter(lambda __lowerCAmelCase : p.requires_grad , model.parameters() ) _UpperCAmelCase : Union[str, Any] = sum([np.prod(p.size() ) for p in model_parameters] ) return params lowerCamelCase__ = logging.getLogger(__name__) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if metric == "rouge2": _UpperCAmelCase : Optional[int] = "{val_avg_rouge2:.4f}-{step_count}" elif metric == "bleu": _UpperCAmelCase : int = "{val_avg_bleu:.4f}-{step_count}" elif metric == "em": _UpperCAmelCase : List[Any] = "{val_avg_em:.4f}-{step_count}" elif metric == "loss": _UpperCAmelCase : Any = "{val_avg_loss:.4f}-{step_count}" else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" " function." ) _UpperCAmelCase : int = ModelCheckpoint( dirpath=__lowerCAmelCase , filename=__lowerCAmelCase , monitor=F"""val_{metric}""" , mode="max" , save_top_k=1 , every_n_epochs=1 , ) return checkpoint_callback def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return EarlyStopping( monitor=F"""val_{metric}""" , mode="min" if "loss" in metric else "max" , patience=__lowerCAmelCase , verbose=__lowerCAmelCase , ) class lowerCAmelCase__ ( pl.Callback ): def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : int , lowerCamelCase__ : Optional[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : List[str] = {F"""lr_group_{i}""": param["lr"] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCamelCase__ ) @rank_zero_only def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : pl.Trainer , lowerCamelCase__ : pl.LightningModule , lowerCamelCase__ : str , lowerCamelCase__ : int=True ) ->None: '''simple docstring''' logger.info(F"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) _UpperCAmelCase : List[Any] = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["log", "progress_bar", "preds"]} ) # Log results _UpperCAmelCase : Optional[int] = Path(pl_module.hparams.output_dir ) if type_path == "test": _UpperCAmelCase : Optional[Any] = od / "test_results.txt" _UpperCAmelCase : Optional[int] = od / "test_generations.txt" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. _UpperCAmelCase : Tuple = od / F"""{type_path}_results/{trainer.global_step:05d}.txt""" _UpperCAmelCase : Union[str, Any] = od / F"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCamelCase__ ) generations_file.parent.mkdir(exist_ok=lowerCamelCase__ ) with open(lowerCamelCase__ , "a+" ) as writer: for key in sorted(lowerCamelCase__ ): if key in ["log", "progress_bar", "preds"]: continue _UpperCAmelCase : Any = metrics[key] if isinstance(lowerCamelCase__ , torch.Tensor ): _UpperCAmelCase : Optional[int] = val.item() _UpperCAmelCase : Optional[int] = F"""{key}: {val:.6f}\n""" writer.write(lowerCamelCase__ ) if not save_generations: return if "preds" in metrics: _UpperCAmelCase : List[Any] = "\n".join(metrics["preds"] ) generations_file.open("w+" ).write(lowerCamelCase__ ) @rank_zero_only def lowerCAmelCase__ ( self : str , lowerCamelCase__ : Tuple , lowerCamelCase__ : Tuple ) ->Optional[int]: '''simple docstring''' try: _UpperCAmelCase : str = pl_module.model.model.num_parameters() except AttributeError: _UpperCAmelCase : Optional[int] = pl_module.model.num_parameters() _UpperCAmelCase : Union[str, Any] = count_trainable_parameters(lowerCamelCase__ ) # mp stands for million parameters trainer.logger.log_metrics({"n_params": npars, "mp": npars / 1E6, "grad_mp": n_trainable_pars / 1E6} ) @rank_zero_only def lowerCAmelCase__ ( self : str , lowerCamelCase__ : pl.Trainer , lowerCamelCase__ : pl.LightningModule ) ->Tuple: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCamelCase__ , lowerCamelCase__ , "test" ) @rank_zero_only def lowerCAmelCase__ ( self : int , lowerCamelCase__ : pl.Trainer , lowerCamelCase__ : Optional[int] ) ->Optional[int]: '''simple docstring''' save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
360
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): if number < 0: raise ValueError("number must not be negative" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
322
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'microsoft/resnet-50': 'https://huggingface.co/microsoft/resnet-50/blob/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): lowerCAmelCase : int = "resnet" lowerCAmelCase : Union[str, Any] = ["basic", "bottleneck"] def __init__( self : Dict , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : Any=64 , lowerCamelCase__ : Optional[int]=[2_56, 5_12, 10_24, 20_48] , lowerCamelCase__ : int=[3, 4, 6, 3] , lowerCamelCase__ : Dict="bottleneck" , lowerCamelCase__ : Dict="relu" , lowerCamelCase__ : List[Any]=False , lowerCamelCase__ : Any=None , lowerCamelCase__ : int=None , **lowerCamelCase__ : Tuple , ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) _UpperCAmelCase : str = num_channels _UpperCAmelCase : List[str] = embedding_size _UpperCAmelCase : Tuple = hidden_sizes _UpperCAmelCase : Dict = depths _UpperCAmelCase : List[Any] = layer_type _UpperCAmelCase : Optional[int] = hidden_act _UpperCAmelCase : Tuple = downsample_in_first_stage _UpperCAmelCase : str = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(lowerCamelCase__ ) + 1 )] _UpperCAmelCase : Union[str, Any] = get_aligned_output_features_output_indices( out_features=lowerCamelCase__ , out_indices=lowerCamelCase__ , stage_names=self.stage_names ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = version.parse("1.11" ) @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowerCAmelCase__ ( self : str ) ->float: '''simple docstring''' return 1E-3
361
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('Program to check whether a number is a Perfect number or not...') lowerCamelCase__ = int(input('Enter number: ').strip()) print(F'''{number} is {"" if perfect(number) else "not "}a Perfect Number.''')
322
0
'''simple docstring''' import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class lowerCAmelCase__ ( unittest.TestCase ): lowerCAmelCase : Union[str, Any] = MODEL_FOR_MASKED_LM_MAPPING lowerCAmelCase : Optional[int] = TF_MODEL_FOR_MASKED_LM_MAPPING def lowerCAmelCase__ ( self : Any ) ->Dict: '''simple docstring''' super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def lowerCAmelCase__ ( self : List[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="tf" ) _UpperCAmelCase : Optional[Any] = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=6 ) , [ {"sequence": "My name is grouped", "score": 2.1E-05, "token": 3_80_15, "token_str": " grouped"}, {"sequence": "My name is accuser", "score": 2.1E-05, "token": 2_55_06, "token_str": " accuser"}, ] , ) _UpperCAmelCase : Dict = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=6 ) , [ { "sequence": "The largest city in France is grouped", "score": 2.1E-05, "token": 3_80_15, "token_str": " grouped", }, { "sequence": "The largest city in France is accuser", "score": 2.1E-05, "token": 2_55_06, "token_str": " accuser", }, ] , ) _UpperCAmelCase : List[str] = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=6 ) , [ {"sequence": "My name is Clara", "score": 2E-05, "token": 1_36_06, "token_str": " Clara"}, {"sequence": "My name is Patrick", "score": 2E-05, "token": 34_99, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 1.9E-05, "token": 29_41, "token_str": " Te"}, ] , ) @require_torch def lowerCAmelCase__ ( self : Tuple ) ->str: '''simple docstring''' _UpperCAmelCase : int = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , top_k=2 , framework="pt" ) _UpperCAmelCase : List[Any] = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=6 ) , [ {"sequence": "My name is Maul", "score": 2.2E-05, "token": 3_56_76, "token_str": " Maul"}, {"sequence": "My name isELS", "score": 2.2E-05, "token": 1_64_16, "token_str": "ELS"}, ] , ) _UpperCAmelCase : List[str] = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=6 ) , [ { "sequence": "The largest city in France is Maul", "score": 2.2E-05, "token": 3_56_76, "token_str": " Maul", }, {"sequence": "The largest city in France isELS", "score": 2.2E-05, "token": 1_64_16, "token_str": "ELS"}, ] , ) _UpperCAmelCase : int = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=6 ) , [ {"sequence": "My name is Patrick", "score": 2.1E-05, "token": 34_99, "token_str": " Patrick"}, {"sequence": "My name is Te", "score": 2E-05, "token": 29_41, "token_str": " Te"}, {"sequence": "My name is Clara", "score": 2E-05, "token": 1_36_06, "token_str": " Clara"}, ] , ) _UpperCAmelCase : List[Any] = unmasker("My name is <mask> <mask>" , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=6 ) , [ [ { "score": 2.2E-05, "token": 3_56_76, "token_str": " Maul", "sequence": "<s>My name is Maul<mask></s>", }, {"score": 2.2E-05, "token": 1_64_16, "token_str": "ELS", "sequence": "<s>My name isELS<mask></s>"}, ], [ { "score": 2.2E-05, "token": 3_56_76, "token_str": " Maul", "sequence": "<s>My name is<mask> Maul</s>", }, {"score": 2.2E-05, "token": 1_64_16, "token_str": "ELS", "sequence": "<s>My name is<mask>ELS</s>"}, ], ] , ) @require_torch_gpu def lowerCAmelCase__ ( self : Dict ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[str] = pipeline("fill-mask" , model="hf-internal-testing/tiny-random-distilbert" , device=0 , framework="pt" ) # convert model to fp16 pipe.model.half() _UpperCAmelCase : Union[str, Any] = pipe("Paris is the [MASK] of France." ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow @require_torch def lowerCAmelCase__ ( self : str ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : List[str] = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="pt" ) self.run_large_test(lowerCamelCase__ ) @slow @require_tf def lowerCAmelCase__ ( self : str ) ->Any: '''simple docstring''' _UpperCAmelCase : Tuple = pipeline(task="fill-mask" , model="distilroberta-base" , top_k=2 , framework="tf" ) self.run_large_test(lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = unmasker("My name is <mask>" ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ {"sequence": "My name is John", "score": 0.0_0_8, "token": 6_10, "token_str": " John"}, {"sequence": "My name is Chris", "score": 0.0_0_7, "token": 15_73, "token_str": " Chris"}, ] , ) _UpperCAmelCase : Union[str, Any] = unmasker("The largest city in France is <mask>" ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ { "sequence": "The largest city in France is Paris", "score": 0.2_5_1, "token": 22_01, "token_str": " Paris", }, { "sequence": "The largest city in France is Lyon", "score": 0.2_1_4, "token": 1_27_90, "token_str": " Lyon", }, ] , ) _UpperCAmelCase : List[str] = unmasker("My name is <mask>" , targets=[" Patrick", " Clara", " Teven"] , top_k=3 ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ {"sequence": "My name is Patrick", "score": 0.0_0_5, "token": 34_99, "token_str": " Patrick"}, {"sequence": "My name is Clara", "score": 0.0_0_0, "token": 1_36_06, "token_str": " Clara"}, {"sequence": "My name is Te", "score": 0.0_0_0, "token": 29_41, "token_str": " Te"}, ] , ) @require_torch def lowerCAmelCase__ ( self : str ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : List[Any] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="pt" ) _UpperCAmelCase : Dict = None _UpperCAmelCase : Tuple = None self.run_pipeline_test(lowerCamelCase__ , [] ) @require_tf def lowerCAmelCase__ ( self : Optional[Any] ) ->Tuple: '''simple docstring''' _UpperCAmelCase : List[Any] = pipeline(task="fill-mask" , model="sshleifer/tiny-distilroberta-base" , framework="tf" ) _UpperCAmelCase : Union[str, Any] = None _UpperCAmelCase : Tuple = None self.run_pipeline_test(lowerCamelCase__ , [] ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : str , lowerCamelCase__ : int ) ->Dict: '''simple docstring''' if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("The provided tokenizer has no mask token, (probably reformer or wav2vec2)" ) _UpperCAmelCase : Optional[int] = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = [ F"""This is another {tokenizer.mask_token} test""", ] return fill_masker, examples def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = fill_masker.tokenizer _UpperCAmelCase : int = fill_masker.model _UpperCAmelCase : str = fill_masker( F"""This is a {tokenizer.mask_token}""" , ) self.assertEqual( lowerCamelCase__ , [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ] , ) _UpperCAmelCase : str = fill_masker([F"""This is a {tokenizer.mask_token}"""] ) self.assertEqual( lowerCamelCase__ , [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ] , ) _UpperCAmelCase : Optional[Any] = fill_masker([F"""This is a {tokenizer.mask_token}""", F"""Another {tokenizer.mask_token} great test."""] ) self.assertEqual( lowerCamelCase__ , [ [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ], [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ], ] , ) with self.assertRaises(lowerCamelCase__ ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(lowerCamelCase__ ): fill_masker("This is" ) self.run_test_top_k(lowerCamelCase__ , lowerCamelCase__ ) self.run_test_targets(lowerCamelCase__ , lowerCamelCase__ ) self.run_test_top_k_targets(lowerCamelCase__ , lowerCamelCase__ ) self.fill_mask_with_duplicate_targets_and_top_k(lowerCamelCase__ , lowerCamelCase__ ) self.fill_mask_with_multiple_masks(lowerCamelCase__ , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : Optional[Any] = tokenizer.get_vocab() _UpperCAmelCase : Any = sorted(vocab.keys() )[:2] # Pipeline argument _UpperCAmelCase : str = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ , targets=lowerCamelCase__ ) _UpperCAmelCase : Tuple = fill_masker(F"""This is a {tokenizer.mask_token}""" ) self.assertEqual( lowerCamelCase__ , [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ] , ) _UpperCAmelCase : Union[str, Any] = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowerCamelCase__ ) _UpperCAmelCase : str = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowerCamelCase__ ) ) # Call argument _UpperCAmelCase : Union[str, Any] = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = fill_masker(F"""This is a {tokenizer.mask_token}""" , targets=lowerCamelCase__ ) self.assertEqual( lowerCamelCase__ , [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ] , ) _UpperCAmelCase : Optional[Any] = {vocab[el] for el in targets} self.assertEqual({el["token"] for el in outputs} , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["token_str"] for el in outputs} , set(lowerCamelCase__ ) ) # Score equivalence _UpperCAmelCase : Union[str, Any] = fill_masker(F"""This is a {tokenizer.mask_token}""" , targets=lowerCamelCase__ ) _UpperCAmelCase : List[str] = [top_mask["token_str"] for top_mask in outputs] _UpperCAmelCase : str = [top_mask["score"] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowerCamelCase__ ) == set(lowerCamelCase__ ): _UpperCAmelCase : Dict = fill_masker(F"""This is a {tokenizer.mask_token}""" , targets=lowerCamelCase__ ) _UpperCAmelCase : List[Any] = [top_mask["score"] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowerCamelCase__ ) , nested_simplify(lowerCamelCase__ ) ) # Raises with invalid with self.assertRaises(lowerCamelCase__ ): _UpperCAmelCase : Any = fill_masker(F"""This is a {tokenizer.mask_token}""" , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowerCamelCase__ ): _UpperCAmelCase : Union[str, Any] = fill_masker(F"""This is a {tokenizer.mask_token}""" , targets=[""] ) with self.assertRaises(lowerCamelCase__ ): _UpperCAmelCase : Optional[int] = fill_masker(F"""This is a {tokenizer.mask_token}""" , targets="" ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Optional[int] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ , top_k=2 ) _UpperCAmelCase : List[Any] = fill_masker(F"""This is a {tokenizer.mask_token}""" ) self.assertEqual( lowerCamelCase__ , [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ] , ) _UpperCAmelCase : str = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) _UpperCAmelCase : Dict = fill_masker(F"""This is a {tokenizer.mask_token}""" , top_k=2 ) self.assertEqual( lowerCamelCase__ , [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ] , ) self.assertEqual(nested_simplify(lowerCamelCase__ ) , nested_simplify(lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : Dict ) ->str: '''simple docstring''' _UpperCAmelCase : Any = tokenizer.get_vocab() _UpperCAmelCase : Tuple = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) # top_k=2, ntargets=3 _UpperCAmelCase : List[Any] = sorted(vocab.keys() )[:3] _UpperCAmelCase : List[str] = fill_masker(F"""This is a {tokenizer.mask_token}""" , top_k=2 , targets=lowerCamelCase__ ) # If we use the most probably targets, and filter differently, we should still # have the same results _UpperCAmelCase : Any = [el["token_str"] for el in sorted(lowerCamelCase__ , key=lambda lowerCamelCase__ : x["score"] , reverse=lowerCamelCase__ )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowerCamelCase__ ).issubset(lowerCamelCase__ ): _UpperCAmelCase : Optional[int] = fill_masker(F"""This is a {tokenizer.mask_token}""" , top_k=3 , targets=lowerCamelCase__ ) # They should yield exactly the same result self.assertEqual(nested_simplify(lowerCamelCase__ ) , nested_simplify(lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : List[Any] , lowerCamelCase__ : str ) ->str: '''simple docstring''' _UpperCAmelCase : List[str] = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) _UpperCAmelCase : List[str] = tokenizer.get_vocab() # String duplicates + id duplicates _UpperCAmelCase : Dict = sorted(vocab.keys() )[:3] _UpperCAmelCase : str = [targets[0], targets[1], targets[0], targets[2], targets[1]] _UpperCAmelCase : int = fill_masker(F"""My name is {tokenizer.mask_token}""" , targets=lowerCamelCase__ , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowerCamelCase__ ) , 3 ) def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Tuple , lowerCamelCase__ : Optional[int] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = FillMaskPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = fill_masker( F"""This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}""" , top_k=2 ) self.assertEqual( lowerCamelCase__ , [ [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ], [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ], [ {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, {"sequence": ANY(lowerCamelCase__ ), "score": ANY(lowerCamelCase__ ), "token": ANY(lowerCamelCase__ ), "token_str": ANY(lowerCamelCase__ )}, ], ] , )
362
'''simple docstring''' from collections.abc import Sequence def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return sum(c * (x**i) for i, c in enumerate(__lowerCAmelCase ) ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Dict = 0.0 for coeff in reversed(__lowerCAmelCase ): _UpperCAmelCase : int = result * x + coeff return result if __name__ == "__main__": lowerCamelCase__ = (0.0, 0.0, 5.0, 9.3, 7.0) lowerCamelCase__ = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
322
0
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class lowerCAmelCase__ : def __init__( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Union[str, Any]=3 , lowerCamelCase__ : Any=7 , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Union[str, Any]=False , lowerCamelCase__ : Union[str, Any]=True , lowerCamelCase__ : Tuple=99 , lowerCamelCase__ : Any=32 , lowerCamelCase__ : Any=5 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : Dict=37 , lowerCamelCase__ : Tuple="gelu" , lowerCamelCase__ : Dict=0.1 , lowerCamelCase__ : Optional[Any]=0.1 , lowerCamelCase__ : Dict=5_12 , lowerCamelCase__ : List[str]=16 , lowerCamelCase__ : List[str]=2 , lowerCamelCase__ : Tuple=0.0_2 , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : str=None , ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Any = parent _UpperCAmelCase : List[str] = batch_size _UpperCAmelCase : Any = seq_length _UpperCAmelCase : Optional[Any] = is_training _UpperCAmelCase : Tuple = use_input_mask _UpperCAmelCase : int = use_token_type_ids _UpperCAmelCase : Optional[Any] = use_labels _UpperCAmelCase : Union[str, Any] = vocab_size _UpperCAmelCase : Dict = hidden_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : str = num_attention_heads _UpperCAmelCase : List[Any] = intermediate_size _UpperCAmelCase : Dict = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Any = attention_probs_dropout_prob _UpperCAmelCase : List[Any] = max_position_embeddings _UpperCAmelCase : List[str] = type_vocab_size _UpperCAmelCase : Optional[int] = type_sequence_label_size _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Union[str, Any] = num_labels _UpperCAmelCase : Optional[int] = num_choices _UpperCAmelCase : str = scope def lowerCAmelCase__ ( self : List[Any] ) ->int: '''simple docstring''' _UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Tuple = None if self.use_input_mask: _UpperCAmelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCAmelCase : Tuple = None _UpperCAmelCase : Optional[Any] = None _UpperCAmelCase : Tuple = None _UpperCAmelCase : Tuple = None if self.use_labels: _UpperCAmelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _UpperCAmelCase : int = ids_tensor([self.batch_size] , self.num_choices ) _UpperCAmelCase : Any = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase__ ( self : Union[str, Any] ) ->int: '''simple docstring''' return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=lowerCamelCase__ , ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Any , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : str , lowerCamelCase__ : List[Any] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : int = FalconModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Tuple = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ ) _UpperCAmelCase : str = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple , lowerCamelCase__ : Any , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : List[str] , lowerCamelCase__ : Tuple , ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Dict = True _UpperCAmelCase : Union[str, Any] = FalconModel(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Union[str, Any] = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , encoder_attention_mask=lowerCamelCase__ , ) _UpperCAmelCase : Optional[Any] = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , ) _UpperCAmelCase : Union[str, Any] = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : str , lowerCamelCase__ : Dict , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : int , ) ->Dict: '''simple docstring''' _UpperCAmelCase : Tuple = FalconForCausalLM(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Union[str, Any] = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : int , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Dict , ) ->str: '''simple docstring''' _UpperCAmelCase : Optional[Any] = True _UpperCAmelCase : List[Any] = True _UpperCAmelCase : Any = FalconForCausalLM(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() # first forward pass _UpperCAmelCase : str = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , encoder_attention_mask=lowerCamelCase__ , use_cache=lowerCamelCase__ , ) _UpperCAmelCase : Tuple = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _UpperCAmelCase : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) _UpperCAmelCase : str = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _UpperCAmelCase : Tuple = torch.cat([input_ids, next_tokens] , dim=-1 ) _UpperCAmelCase : Tuple = torch.cat([input_mask, next_mask] , dim=-1 ) _UpperCAmelCase : Union[str, Any] = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , encoder_attention_mask=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , )["hidden_states"][0] _UpperCAmelCase : List[Any] = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , encoder_attention_mask=lowerCamelCase__ , past_key_values=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , )["hidden_states"][0] # select random slice _UpperCAmelCase : Optional[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() _UpperCAmelCase : Optional[int] = output_from_no_past[:, -3:, random_slice_idx].detach() _UpperCAmelCase : int = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1E-3 ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[str] = self.prepare_config_and_inputs() ( _UpperCAmelCase ) : Optional[Any] = config_and_inputs _UpperCAmelCase : Dict = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Dict = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) lowerCAmelCase : int = (FalconForCausalLM,) if is_torch_available() else () lowerCAmelCase : str = ( { "feature-extraction": FalconModel, "text-classification": FalconForSequenceClassification, "text-generation": FalconForCausalLM, "question-answering": FalconForQuestionAnswering, "token-classification": FalconForTokenClassification, "zero-shot": FalconForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase : List[Any] = False lowerCAmelCase : int = False def lowerCAmelCase__ ( self : List[Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : int = FalconModelTester(self ) _UpperCAmelCase : Dict = ConfigTester(self , config_class=lowerCamelCase__ , hidden_size=37 ) def lowerCAmelCase__ ( self : List[str] ) ->Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCAmelCase__ ( self : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: _UpperCAmelCase : Optional[Any] = alibi self.model_tester.create_and_check_model(lowerCamelCase__ , *lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[Any] ) ->int: '''simple docstring''' _UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Optional[int] = 3 _UpperCAmelCase : Any = input_dict["input_ids"] _UpperCAmelCase : Optional[Any] = input_ids.ne(1 ).to(lowerCamelCase__ ) _UpperCAmelCase : Any = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _UpperCAmelCase : Tuple = FalconForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Tuple = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase__ ( self : Union[str, Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : List[Any] = 3 _UpperCAmelCase : Tuple = "single_label_classification" _UpperCAmelCase : Any = input_dict["input_ids"] _UpperCAmelCase : Dict = input_ids.ne(1 ).to(lowerCamelCase__ ) _UpperCAmelCase : List[str] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _UpperCAmelCase : List[Any] = FalconForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Union[str, Any] = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase__ ( self : int ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Optional[int] = input_dict["input_ids"] _UpperCAmelCase : Dict = FalconForCausalLM(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Union[str, Any] = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _UpperCAmelCase : int = input_ids.shape[0] _UpperCAmelCase : Tuple = model._convert_to_rw_cache(result.past_key_values ) _UpperCAmelCase : int = model._convert_cache_to_standard_format(lowerCamelCase__ , lowerCamelCase__ ) for layer in range(len(lowerCamelCase__ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def lowerCAmelCase__ ( self : str ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Union[str, Any] = 3 _UpperCAmelCase : List[str] = "multi_label_classification" _UpperCAmelCase : List[str] = input_dict["input_ids"] _UpperCAmelCase : Any = input_ids.ne(1 ).to(lowerCamelCase__ ) _UpperCAmelCase : str = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _UpperCAmelCase : List[Any] = FalconForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : int = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCAmelCase__ ( self : str ) ->str: '''simple docstring''' for model_class in self.all_generative_model_classes: _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(lowerCamelCase__ , "use_cache" ): return _UpperCAmelCase : int = model_class(lowerCamelCase__ ).to(lowerCamelCase__ ) if "use_cache" not in inputs: _UpperCAmelCase : str = True _UpperCAmelCase : List[Any] = model(**lowerCamelCase__ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return _UpperCAmelCase : int = ( getattr(lowerCamelCase__ , "decoder_layers" , lowerCamelCase__ ) or getattr(lowerCamelCase__ , "num_decoder_layers" , lowerCamelCase__ ) or config.num_hidden_layers ) _UpperCAmelCase : Optional[Any] = getattr(lowerCamelCase__ , "num_kv_heads" , config.num_attention_heads ) _UpperCAmelCase : Union[str, Any] = getattr(lowerCamelCase__ , "d_model" , config.hidden_size ) _UpperCAmelCase : Tuple = embed_dim // num_attention_heads _UpperCAmelCase : Optional[Any] = outputs["past_key_values"] self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = inputs["input_ids"].shape for i in range(lowerCamelCase__ ): if config.new_decoder_architecture: _UpperCAmelCase : Tuple = config.num_attention_heads elif config.multi_query: _UpperCAmelCase : List[Any] = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : int ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) _UpperCAmelCase : List[Any] = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(lowerCamelCase__ ) _UpperCAmelCase : Tuple = tokenizer("My favorite food is" , return_tensors="pt" ).to(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) _UpperCAmelCase : Dict = model.generate(**lowerCamelCase__ , do_sample=lowerCamelCase__ , max_new_tokens=19 ) _UpperCAmelCase : int = tokenizer.batch_decode(lowerCamelCase__ )[0] self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : List[Any] ) ->List[str]: '''simple docstring''' for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: _UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = FalconForCausalLM.from_pretrained(lowerCamelCase__ ) model.eval() model.to(lowerCamelCase__ ) _UpperCAmelCase : List[str] = tokenizer("My favorite food is" , return_tensors="pt" ).to(lowerCamelCase__ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**lowerCamelCase__ , do_sample=lowerCamelCase__ , max_new_tokens=4 ) model.generate(**lowerCamelCase__ , do_sample=lowerCamelCase__ , max_new_tokens=4 ) model.generate(**lowerCamelCase__ , num_beams=2 , max_new_tokens=4 ) @slow def lowerCAmelCase__ ( self : List[str] ) ->str: '''simple docstring''' with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: _UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : List[str] = FalconForCausalLM.from_pretrained(lowerCamelCase__ ) model.eval() model.to(device=lowerCamelCase__ ) _UpperCAmelCase : Tuple = tokenizer("My favorite food is" , return_tensors="pt" ).to(lowerCamelCase__ ) # Test results are the same with and without cache _UpperCAmelCase : int = model.generate(**lowerCamelCase__ , do_sample=lowerCamelCase__ , max_new_tokens=20 , use_cache=lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = model.generate(**lowerCamelCase__ , do_sample=lowerCamelCase__ , max_new_tokens=20 , use_cache=lowerCamelCase__ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
363
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[Any] = len(__lowerCAmelCase ) _UpperCAmelCase : Tuple = sum(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _UpperCAmelCase : Any = True for i in range(1 , s + 1 ): _UpperCAmelCase : List[Any] = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): _UpperCAmelCase : Optional[int] = dp[i][j - 1] if arr[i - 1] <= j: _UpperCAmelCase : Any = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: _UpperCAmelCase : List[Any] = s - 2 * j break return diff
322
0
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def __lowerCAmelCase (__lowerCAmelCase ): # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0x4_e_0_0 and cp <= 0x9_f_f_f) or (cp >= 0x3_4_0_0 and cp <= 0x4_d_b_f) # or (cp >= 0x2_0_0_0_0 and cp <= 0x2_a_6_d_f) # or (cp >= 0x2_a_7_0_0 and cp <= 0x2_b_7_3_f) # or (cp >= 0x2_b_7_4_0 and cp <= 0x2_b_8_1_f) # or (cp >= 0x2_b_8_2_0 and cp <= 0x2_c_e_a_f) # or (cp >= 0xf_9_0_0 and cp <= 0xf_a_f_f) or (cp >= 0x2_f_8_0_0 and cp <= 0x2_f_a_1_f) # ): # return True return False def __lowerCAmelCase (__lowerCAmelCase ): # word like '180' or '身高' or '神' for char in word: _UpperCAmelCase : Union[str, Any] = ord(__lowerCAmelCase ) if not _is_chinese_char(__lowerCAmelCase ): return 0 return 1 def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Tuple = set() for token in tokens: _UpperCAmelCase : Union[str, Any] = len(__lowerCAmelCase ) > 1 and is_chinese(__lowerCAmelCase ) if chinese_word: word_set.add(__lowerCAmelCase ) _UpperCAmelCase : str = list(__lowerCAmelCase ) return word_list def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if not chinese_word_set: return bert_tokens _UpperCAmelCase : int = max([len(__lowerCAmelCase ) for w in chinese_word_set] ) _UpperCAmelCase : Optional[int] = bert_tokens _UpperCAmelCase : Dict = 0, len(__lowerCAmelCase ) while start < end: _UpperCAmelCase : List[Any] = True if is_chinese(bert_word[start] ): _UpperCAmelCase : Optional[Any] = min(end - start , __lowerCAmelCase ) for i in range(__lowerCAmelCase , 1 , -1 ): _UpperCAmelCase : Tuple = "".join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _UpperCAmelCase : Any = "##" + bert_word[j] _UpperCAmelCase : List[Any] = start + i _UpperCAmelCase : Dict = False break if single_word: start += 1 return bert_word def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [] for i in range(0 , len(__lowerCAmelCase ) , 100 ): _UpperCAmelCase : Union[str, Any] = ltp_tokenizer.pipeline(lines[i : i + 100] , tasks=["cws"] ).cws _UpperCAmelCase : Union[str, Any] = [get_chinese_word(__lowerCAmelCase ) for r in res] ltp_res.extend(__lowerCAmelCase ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = [] for i in range(0 , len(__lowerCAmelCase ) , 100 ): _UpperCAmelCase : List[Any] = bert_tokenizer(lines[i : i + 100] , add_special_tokens=__lowerCAmelCase , truncation=__lowerCAmelCase , max_length=512 ) bert_res.extend(res["input_ids"] ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = [] for input_ids, chinese_word in zip(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [] for id in input_ids: _UpperCAmelCase : Optional[Any] = bert_tokenizer._convert_id_to_token(__lowerCAmelCase ) input_tokens.append(__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = add_sub_symbol(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Dict = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__lowerCAmelCase ): if token[:2] == "##": _UpperCAmelCase : Optional[Any] = token[2:] # save chinese tokens' pos if len(__lowerCAmelCase ) == 1 and _is_chinese_char(ord(__lowerCAmelCase ) ): ref_id.append(__lowerCAmelCase ) ref_ids.append(__lowerCAmelCase ) assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ) return ref_ids def __lowerCAmelCase (__lowerCAmelCase ): # For Chinese (Ro)Bert, the best result is from : RoBERTa-wwm-ext (https://github.com/ymcui/Chinese-BERT-wwm) # If we want to fine-tune these model, we have to use same tokenizer : LTP (https://github.com/HIT-SCIR/ltp) with open(args.file_name , "r" , encoding="utf-8" ) as f: _UpperCAmelCase : Optional[int] = f.readlines() _UpperCAmelCase : Union[str, Any] = [line.strip() for line in data if len(__lowerCAmelCase ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _UpperCAmelCase : str = LTP(args.ltp ) # faster in GPU device _UpperCAmelCase : List[Any] = BertTokenizer.from_pretrained(args.bert ) _UpperCAmelCase : Any = prepare_ref(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) with open(args.save_path , "w" , encoding="utf-8" ) as f: _UpperCAmelCase : Optional[int] = [json.dumps(__lowerCAmelCase ) + "\n" for ref in ref_ids] f.writelines(__lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser(description='prepare_chinese_ref') parser.add_argument( '--file_name', required=False, type=str, default='./resources/chinese-demo.txt', help='file need process, same as training data in lm', ) parser.add_argument( '--ltp', required=False, type=str, default='./resources/ltp', help='resources for LTP tokenizer, usually a path', ) parser.add_argument( '--bert', required=False, type=str, default='./resources/robert', help='resources for Bert tokenizer', ) parser.add_argument( '--save_path', required=False, type=str, default='./resources/ref.txt', help='path to save res', ) lowerCamelCase__ = parser.parse_args() main(args)
364
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'microsoft/resnet-50': 'https://huggingface.co/microsoft/resnet-50/blob/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): lowerCAmelCase : int = "resnet" lowerCAmelCase : Union[str, Any] = ["basic", "bottleneck"] def __init__( self : Dict , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : Any=64 , lowerCamelCase__ : Optional[int]=[2_56, 5_12, 10_24, 20_48] , lowerCamelCase__ : int=[3, 4, 6, 3] , lowerCamelCase__ : Dict="bottleneck" , lowerCamelCase__ : Dict="relu" , lowerCamelCase__ : List[Any]=False , lowerCamelCase__ : Any=None , lowerCamelCase__ : int=None , **lowerCamelCase__ : Tuple , ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) _UpperCAmelCase : str = num_channels _UpperCAmelCase : List[str] = embedding_size _UpperCAmelCase : Tuple = hidden_sizes _UpperCAmelCase : Dict = depths _UpperCAmelCase : List[Any] = layer_type _UpperCAmelCase : Optional[int] = hidden_act _UpperCAmelCase : Tuple = downsample_in_first_stage _UpperCAmelCase : str = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(lowerCamelCase__ ) + 1 )] _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = get_aligned_output_features_output_indices( out_features=lowerCamelCase__ , out_indices=lowerCamelCase__ , stage_names=self.stage_names ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = version.parse("1.11" ) @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowerCAmelCase__ ( self : str ) ->float: '''simple docstring''' return 1E-3
322
0
'''simple docstring''' import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) lowerCamelCase__ = logging.getLogger() def __lowerCAmelCase (): _UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("-f" ) _UpperCAmelCase : int = parser.parse_args() return args.f class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : str ) ->None: '''simple docstring''' _UpperCAmelCase : Dict = logging.StreamHandler(sys.stdout ) logger.addHandler(lowerCamelCase__ ) def lowerCAmelCase__ ( self : int , lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : int = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowerCamelCase__ , "argv" , lowerCamelCase__ ): _UpperCAmelCase : Any = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowerCamelCase__ , 0.6_6_6 ) @slow @require_torch_non_multi_gpu def lowerCAmelCase__ ( self : str ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase__ ) _UpperCAmelCase : Dict = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase__ )
365
'''simple docstring''' from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowerCamelCase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __lowerCAmelCase (__lowerCAmelCase ): if isinstance(__lowerCAmelCase , torch.Tensor ): return image elif isinstance(__lowerCAmelCase , PIL.Image.Image ): _UpperCAmelCase : int = [image] _UpperCAmelCase : str = [trans(img.convert("RGB" ) ) for img in image] _UpperCAmelCase : Optional[Any] = torch.stack(__lowerCAmelCase ) return image class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Tuple , lowerCamelCase__ : int , lowerCamelCase__ : int ) ->int: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM _UpperCAmelCase : Tuple = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : str ) ->Union[str, Any]: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(F"""The value of strength should in [0.0, 1.0] but is {strength}""" ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Dict , lowerCamelCase__ : List[str] , lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = min(int(num_inference_steps * strength ) , lowerCamelCase__ ) _UpperCAmelCase : str = max(num_inference_steps - init_timestep , 0 ) _UpperCAmelCase : List[str] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : str , lowerCamelCase__ : str , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[Any]=None ) ->str: '''simple docstring''' if not isinstance(lowerCamelCase__ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowerCamelCase__ )}""" ) _UpperCAmelCase : Union[str, Any] = image.to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(lowerCamelCase__ )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) _UpperCAmelCase : List[str] = init_latents.shape _UpperCAmelCase : Optional[int] = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=lowerCamelCase__ , dtype=lowerCamelCase__ ) # get latents print("add noise to latents at timestep" , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.scheduler.add_noise(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = init_latents return latents @torch.no_grad() def __call__( self : Any , lowerCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image] = None , lowerCamelCase__ : float = 0.8 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase__ : float = 0.0 , lowerCamelCase__ : int = 50 , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , ) ->Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowerCamelCase__ ) # 2. Preprocess image _UpperCAmelCase : Dict = preprocess(lowerCamelCase__ ) # 3. set timesteps self.scheduler.set_timesteps(lowerCamelCase__ , device=self.device ) _UpperCAmelCase , _UpperCAmelCase : Any = self.get_timesteps(lowerCamelCase__ , lowerCamelCase__ , self.device ) _UpperCAmelCase : List[Any] = timesteps[:1].repeat(lowerCamelCase__ ) # 4. Prepare latent variables _UpperCAmelCase : Optional[int] = self.prepare_latents(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , self.unet.dtype , self.device , lowerCamelCase__ ) _UpperCAmelCase : Any = latents # 5. Denoising loop for t in self.progress_bar(lowerCamelCase__ ): # 1. predict noise model_output _UpperCAmelCase : Union[str, Any] = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _UpperCAmelCase : int = self.scheduler.step( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , eta=lowerCamelCase__ , use_clipped_model_output=lowerCamelCase__ , generator=lowerCamelCase__ , ).prev_sample _UpperCAmelCase : Dict = (image / 2 + 0.5).clamp(0 , 1 ) _UpperCAmelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCAmelCase : str = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowerCamelCase__ )
322
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : int = "speech_to_text_2" lowerCAmelCase : str = ["past_key_values"] lowerCAmelCase : int = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[Any] , lowerCamelCase__ : Tuple=1_00_00 , lowerCamelCase__ : Any=6 , lowerCamelCase__ : Tuple=20_48 , lowerCamelCase__ : List[Any]=4 , lowerCamelCase__ : Tuple=0.0 , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Tuple="relu" , lowerCamelCase__ : Dict=2_56 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : List[Any]=0.0 , lowerCamelCase__ : Optional[int]=0.0 , lowerCamelCase__ : List[Any]=0.0_2 , lowerCamelCase__ : Tuple=2 , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Any=1 , lowerCamelCase__ : int=0 , lowerCamelCase__ : str=2 , lowerCamelCase__ : List[Any]=10_24 , **lowerCamelCase__ : str , ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = vocab_size _UpperCAmelCase : Optional[int] = d_model _UpperCAmelCase : List[Any] = decoder_ffn_dim _UpperCAmelCase : Any = decoder_layers _UpperCAmelCase : int = decoder_attention_heads _UpperCAmelCase : Any = dropout _UpperCAmelCase : List[Any] = attention_dropout _UpperCAmelCase : Optional[int] = activation_dropout _UpperCAmelCase : List[Any] = activation_function _UpperCAmelCase : int = init_std _UpperCAmelCase : Dict = decoder_layerdrop _UpperCAmelCase : str = use_cache _UpperCAmelCase : Union[str, Any] = decoder_layers _UpperCAmelCase : Optional[Any] = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCAmelCase : Any = max_target_positions super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
366
'''simple docstring''' from __future__ import annotations from collections.abc import Callable lowerCamelCase__ = list[list[float | int]] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(size + 1 )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : float for row in range(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = matrix[row][col] _UpperCAmelCase : Optional[int] = vector[row][0] _UpperCAmelCase : int = 0 _UpperCAmelCase : Union[str, Any] = 0 while row < size and col < size: # pivoting _UpperCAmelCase : Optional[Any] = max((abs(augmented[rowa][col] ), rowa) for rowa in range(__lowerCAmelCase , __lowerCAmelCase ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: _UpperCAmelCase , _UpperCAmelCase : str = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , __lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[rowa][col] / augmented[row][col] _UpperCAmelCase : Optional[Any] = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , __lowerCAmelCase ): for row in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[row][col] / augmented[col][col] for cola in range(__lowerCAmelCase , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(__lowerCAmelCase ) ] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(__lowerCAmelCase )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix = [[0] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int for x_val, y_val in enumerate(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = (x_val + 1) ** (size - col - 1) _UpperCAmelCase : int = y_val _UpperCAmelCase : List[str] = solve(__lowerCAmelCase , __lowerCAmelCase ) def interpolated_func(__lowerCAmelCase ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(__lowerCAmelCase ) ) return interpolated_func def __lowerCAmelCase (__lowerCAmelCase ): return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def __lowerCAmelCase (__lowerCAmelCase = question_function , __lowerCAmelCase = 10 ): _UpperCAmelCase : list[int] = [func(__lowerCAmelCase ) for x_val in range(1 , order + 1 )] _UpperCAmelCase : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] _UpperCAmelCase : int = 0 _UpperCAmelCase : Callable[[int], int] _UpperCAmelCase : int for poly in polynomials: _UpperCAmelCase : int = 1 while func(__lowerCAmelCase ) == poly(__lowerCAmelCase ): x_val += 1 ret += poly(__lowerCAmelCase ) return ret if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'Intel/dpt-large': 'https://huggingface.co/Intel/dpt-large/resolve/main/config.json', # See all DPT models at https://huggingface.co/models?filter=dpt } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = "dpt" def __init__( self : int , lowerCamelCase__ : int=7_68 , lowerCamelCase__ : Any=12 , lowerCamelCase__ : str=12 , lowerCamelCase__ : Dict=30_72 , lowerCamelCase__ : Any="gelu" , lowerCamelCase__ : int=0.0 , lowerCamelCase__ : Optional[int]=0.0 , lowerCamelCase__ : str=0.0_2 , lowerCamelCase__ : int=1E-12 , lowerCamelCase__ : Optional[Any]=3_84 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : Dict=3 , lowerCamelCase__ : Tuple=False , lowerCamelCase__ : Union[str, Any]=True , lowerCamelCase__ : str=[2, 5, 8, 11] , lowerCamelCase__ : Tuple="project" , lowerCamelCase__ : List[str]=[4, 2, 1, 0.5] , lowerCamelCase__ : Optional[Any]=[96, 1_92, 3_84, 7_68] , lowerCamelCase__ : int=2_56 , lowerCamelCase__ : Optional[int]=-1 , lowerCamelCase__ : Any=False , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[Any]=0.4 , lowerCamelCase__ : str=2_55 , lowerCamelCase__ : int=0.1 , lowerCamelCase__ : str=[1, 10_24, 24, 24] , lowerCamelCase__ : List[str]=[0, 1] , lowerCamelCase__ : Optional[int]=None , **lowerCamelCase__ : List[Any] , ) ->str: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Tuple = hidden_size _UpperCAmelCase : Tuple = is_hybrid if self.is_hybrid: if backbone_config is None: logger.info("Initializing the config with a `BiT` backbone." ) _UpperCAmelCase : List[Any] = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, } _UpperCAmelCase : str = BitConfig(**lowerCamelCase__ ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): logger.info("Initializing the config with a `BiT` backbone." ) _UpperCAmelCase : int = BitConfig(**lowerCamelCase__ ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Any = backbone_config else: raise ValueError( F"""backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}.""" ) _UpperCAmelCase : str = backbone_featmap_shape _UpperCAmelCase : int = neck_ignore_stages if readout_type != "project": raise ValueError("Readout type must be 'project' when using `DPT-hybrid` mode." ) else: _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : int = None _UpperCAmelCase : Any = [] _UpperCAmelCase : Tuple = num_hidden_layers _UpperCAmelCase : str = num_attention_heads _UpperCAmelCase : str = intermediate_size _UpperCAmelCase : Optional[Any] = hidden_act _UpperCAmelCase : Optional[int] = hidden_dropout_prob _UpperCAmelCase : List[Any] = attention_probs_dropout_prob _UpperCAmelCase : Optional[Any] = initializer_range _UpperCAmelCase : List[Any] = layer_norm_eps _UpperCAmelCase : str = image_size _UpperCAmelCase : Dict = patch_size _UpperCAmelCase : Tuple = num_channels _UpperCAmelCase : int = qkv_bias _UpperCAmelCase : Union[str, Any] = backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError("Readout_type must be one of ['ignore', 'add', 'project']" ) _UpperCAmelCase : Optional[Any] = readout_type _UpperCAmelCase : str = reassemble_factors _UpperCAmelCase : List[str] = neck_hidden_sizes _UpperCAmelCase : List[str] = fusion_hidden_size _UpperCAmelCase : List[Any] = head_in_index _UpperCAmelCase : Optional[int] = use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) _UpperCAmelCase : List[Any] = use_auxiliary_head _UpperCAmelCase : List[str] = auxiliary_loss_weight _UpperCAmelCase : Dict = semantic_loss_ignore_index _UpperCAmelCase : Optional[int] = semantic_classifier_dropout def lowerCAmelCase__ ( self : Any ) ->Any: '''simple docstring''' _UpperCAmelCase : Optional[int] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: _UpperCAmelCase : int = self.backbone_config.to_dict() _UpperCAmelCase : Optional[int] = self.__class__.model_type return output
367
'''simple docstring''' from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
322
0
'''simple docstring''' from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : int = "new-model" if is_tf_available(): class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : int = NewModelConfig @require_tf class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : str ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[int] = "bert-base-cased" _UpperCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[str] = TFAutoModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : Union[str, Any] ) ->int: '''simple docstring''' _UpperCAmelCase : Tuple = "bert-base-cased" _UpperCAmelCase : List[Any] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Tuple = TFAutoModelForPreTraining.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : Tuple ) ->Tuple: '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase : Optional[Any] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Tuple = TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase : Any = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : List[Any] ) ->Optional[Any]: '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase : str = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Dict = TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : List[Any] ) ->Dict: '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase : List[str] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : str ) ->Optional[int]: '''simple docstring''' for model_name in ["bert-base-uncased"]: _UpperCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[str] = TFAutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : Optional[Any] ) ->int: '''simple docstring''' for model_name in ["bert-base-uncased"]: _UpperCAmelCase : int = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = TFAutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow @require_tensorflow_probability def lowerCAmelCase__ ( self : Any ) ->Optional[Any]: '''simple docstring''' for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _UpperCAmelCase : List[str] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Dict = TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 1_44_10 ) def lowerCAmelCase__ ( self : Dict ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Dict = TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 1_44_10 ) def lowerCAmelCase__ ( self : str ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[Any] = TFAutoModel.from_pretrained("sgugger/funnel-random-tiny" ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = copy.deepcopy(model.config ) _UpperCAmelCase : List[Any] = ["FunnelBaseModel"] _UpperCAmelCase : Union[str, Any] = TFAutoModel.from_config(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = TFAutoModel.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' try: AutoConfig.register("new-model" , lowerCamelCase__ ) _UpperCAmelCase : List[str] = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowerCamelCase__ ): auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCamelCase__ ): auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) # Now that the config is registered, it can be used as any other config with the auto-API _UpperCAmelCase : Tuple = BertModelTester(self ).get_config() _UpperCAmelCase : Union[str, Any] = NewModelConfig(**tiny_config.to_dict() ) _UpperCAmelCase : Tuple = auto_class.from_config(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : int = auto_class.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def lowerCAmelCase__ ( self : Union[str, Any] ) ->Tuple: '''simple docstring''' with self.assertRaisesRegex( lowerCamelCase__ , "bert-base is not a local folder and is not a valid model identifier" ): _UpperCAmelCase : Any = TFAutoModel.from_pretrained("bert-base" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Any: '''simple docstring''' with self.assertRaisesRegex( lowerCamelCase__ , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): _UpperCAmelCase : Union[str, Any] = TFAutoModel.from_pretrained(lowerCamelCase__ , revision="aaaaaa" ) def lowerCAmelCase__ ( self : List[Any] ) ->Union[str, Any]: '''simple docstring''' with self.assertRaisesRegex( lowerCamelCase__ , "hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin" , ): _UpperCAmelCase : Optional[int] = TFAutoModel.from_pretrained("hf-internal-testing/config-no-model" ) def lowerCAmelCase__ ( self : int ) ->int: '''simple docstring''' with self.assertRaisesRegex(lowerCamelCase__ , "Use `from_pt=True` to load this model" ): _UpperCAmelCase : str = TFAutoModel.from_pretrained("hf-internal-testing/tiny-bert-pt-only" ) def lowerCAmelCase__ ( self : Any ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Optional[int] = TFAutoModel.from_pretrained("hf-internal-testing/tiny-random-bert" ) with RequestCounter() as counter: _UpperCAmelCase : Tuple = TFAutoModel.from_pretrained("hf-internal-testing/tiny-random-bert" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _UpperCAmelCase : str = TFAutoModel.from_pretrained("ArthurZ/tiny-random-bert-sharded" ) with RequestCounter() as counter: _UpperCAmelCase : str = TFAutoModel.from_pretrained("ArthurZ/tiny-random-bert-sharded" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
368
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar lowerCamelCase__ = TypeVar('T') class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Union[str, Any] , lowerCamelCase__ : T ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Dict = data _UpperCAmelCase : Node[T] | None = None def __str__( self : Any ) ->str: '''simple docstring''' return F"""{self.data}""" class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Tuple ) ->None: '''simple docstring''' _UpperCAmelCase : Node[T] | None = None def __iter__( self : List[str] ) ->Iterator[T]: '''simple docstring''' _UpperCAmelCase : Any = self.top while node: yield node.data _UpperCAmelCase : Dict = node.next def __str__( self : Dict ) ->str: '''simple docstring''' return "->".join([str(lowerCamelCase__ ) for item in self] ) def __len__( self : Optional[int] ) ->int: '''simple docstring''' return len(tuple(iter(self ) ) ) def lowerCAmelCase__ ( self : List[Any] ) ->bool: '''simple docstring''' return self.top is None def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : T ) ->None: '''simple docstring''' _UpperCAmelCase : List[Any] = Node(lowerCamelCase__ ) if not self.is_empty(): _UpperCAmelCase : Tuple = self.top _UpperCAmelCase : List[str] = node def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("pop from empty stack" ) assert isinstance(self.top , lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.top _UpperCAmelCase : Optional[Any] = self.top.next return pop_node.data def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("peek from empty stack" ) assert self.top is not None return self.top.data def lowerCAmelCase__ ( self : List[Any] ) ->None: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = None if __name__ == "__main__": from doctest import testmod testmod()
322
0
'''simple docstring''' import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS logging.basicConfig(level=logging.DEBUG) lowerCamelCase__ = logging.getLogger() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Tuple = "\n".join(__lowerCAmelCase ) Path(__lowerCAmelCase ).open("w" ).writelines(__lowerCAmelCase ) lowerCamelCase__ = 'patrickvonplaten/t5-tiny-random' lowerCamelCase__ = 'sshleifer/bart-tiny-random' lowerCamelCase__ = 'sshleifer/tiny-mbart' lowerCamelCase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) logging.disable(logging.CRITICAL) # remove noisy download output from tracebacks class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Any ) ->Tuple: '''simple docstring''' _UpperCAmelCase : List[str] = Path(self.get_auto_remove_tmp_dir() ) / "utest_input.source" _UpperCAmelCase : Dict = input_file_name.parent / "utest_output.txt" assert not output_file_name.exists() _UpperCAmelCase : List[str] = [" New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County."] _dump_articles(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Dict = str(Path(self.get_auto_remove_tmp_dir() ) / "scores.json" ) _UpperCAmelCase : Optional[Any] = "translation_en_to_de" if model == T5_TINY else "summarization" _UpperCAmelCase : Any = F""" run_eval_search.py {model} {input_file_name} {output_file_name} --score_path {score_path} --task {task} --num_beams 2 --length_penalty 2.0 """.split() with patch.object(lowerCamelCase__ , "argv" , lowerCamelCase__ ): run_generate() assert Path(lowerCamelCase__ ).exists() # os.remove(Path(output_file_name)) def lowerCAmelCase__ ( self : List[str] ) ->Optional[Any]: '''simple docstring''' self.run_eval_tester(lowerCamelCase__ ) @parameterized.expand([BART_TINY, MBART_TINY] ) @slow def lowerCAmelCase__ ( self : int , lowerCamelCase__ : List[Any] ) ->Dict: '''simple docstring''' self.run_eval_tester(lowerCamelCase__ ) @parameterized.expand([T5_TINY, MBART_TINY] ) @slow def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : List[str] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = Path(self.get_auto_remove_tmp_dir() ) / "utest_input.source" _UpperCAmelCase : Optional[int] = input_file_name.parent / "utest_output.txt" assert not output_file_name.exists() _UpperCAmelCase : int = { "en": ["Machine learning is great, isn't it?", "I like to eat bananas", "Tomorrow is another great day!"], "de": [ "Maschinelles Lernen ist großartig, oder?", "Ich esse gerne Bananen", "Morgen ist wieder ein toller Tag!", ], } _UpperCAmelCase : Tuple = Path(self.get_auto_remove_tmp_dir() ) _UpperCAmelCase : Any = str(tmp_dir / "scores.json" ) _UpperCAmelCase : List[Any] = str(tmp_dir / "val.target" ) _dump_articles(lowerCamelCase__ , text["en"] ) _dump_articles(lowerCamelCase__ , text["de"] ) _UpperCAmelCase : List[str] = "translation_en_to_de" if model == T5_TINY else "summarization" _UpperCAmelCase : Optional[int] = F""" run_eval_search.py {model} {str(lowerCamelCase__ )} {str(lowerCamelCase__ )} --score_path {score_path} --reference_path {reference_path} --task {task} """.split() testargs.extend(["--search", "num_beams=1:2 length_penalty=0.9:1.0"] ) with patch.object(lowerCamelCase__ , "argv" , lowerCamelCase__ ): with CaptureStdout() as cs: run_search() _UpperCAmelCase : Dict = [" num_beams | length_penalty", model, "Best score args"] _UpperCAmelCase : int = ["Info"] if "translation" in task: expected_strings.append("bleu" ) else: expected_strings.extend(lowerCamelCase__ ) for w in expected_strings: assert w in cs.out for w in un_expected_strings: assert w not in cs.out assert Path(lowerCamelCase__ ).exists() os.remove(Path(lowerCamelCase__ ) )
369
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : int = "speech_to_text_2" lowerCAmelCase : str = ["past_key_values"] lowerCAmelCase : int = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[Any] , lowerCamelCase__ : Tuple=1_00_00 , lowerCamelCase__ : Any=6 , lowerCamelCase__ : Tuple=20_48 , lowerCamelCase__ : List[Any]=4 , lowerCamelCase__ : Tuple=0.0 , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Tuple="relu" , lowerCamelCase__ : Dict=2_56 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : List[Any]=0.0 , lowerCamelCase__ : Optional[int]=0.0 , lowerCamelCase__ : List[Any]=0.0_2 , lowerCamelCase__ : Tuple=2 , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Any=1 , lowerCamelCase__ : int=0 , lowerCamelCase__ : str=2 , lowerCamelCase__ : List[Any]=10_24 , **lowerCamelCase__ : str , ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = vocab_size _UpperCAmelCase : Optional[int] = d_model _UpperCAmelCase : List[Any] = decoder_ffn_dim _UpperCAmelCase : Any = decoder_layers _UpperCAmelCase : int = decoder_attention_heads _UpperCAmelCase : Any = dropout _UpperCAmelCase : List[Any] = attention_dropout _UpperCAmelCase : Optional[int] = activation_dropout _UpperCAmelCase : List[Any] = activation_function _UpperCAmelCase : int = init_std _UpperCAmelCase : Dict = decoder_layerdrop _UpperCAmelCase : str = use_cache _UpperCAmelCase : Union[str, Any] = decoder_layers _UpperCAmelCase : Optional[Any] = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCAmelCase : Any = max_target_positions super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
322
0
'''simple docstring''' from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets lowerCamelCase__ = '\\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n' lowerCamelCase__ = '\\nGLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n' lowerCamelCase__ = '\nCompute GLUE evaluation metric associated to each GLUE dataset.\nArgs:\n predictions: list of predictions to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\nReturns: depending on the GLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "pearson": Pearson Correlation\n "spearmanr": Spearman Correlation\n "matthews_correlation": Matthew Correlation\nExamples:\n\n >>> glue_metric = datasets.load_metric(\'glue\', \'sst2\') # \'sst2\' or any of ["mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> glue_metric = datasets.load_metric(\'glue\', \'mrpc\') # \'mrpc\' or \'qqp\'\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> glue_metric = datasets.load_metric(\'glue\', \'stsb\')\n >>> references = [0., 1., 2., 3., 4., 5.]\n >>> predictions = [0., 1., 2., 3., 4., 5.]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print({"pearson": round(results["pearson"], 2), "spearmanr": round(results["spearmanr"], 2)})\n {\'pearson\': 1.0, \'spearmanr\': 1.0}\n\n >>> glue_metric = datasets.load_metric(\'glue\', \'cola\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'matthews_correlation\': 1.0}\n' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return float((preds == labels).mean() ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Tuple = simple_accuracy(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Optional[int] = float(fa_score(y_true=__lowerCAmelCase , y_pred=__lowerCAmelCase ) ) return { "accuracy": acc, "f1": fa, } def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : str = float(pearsonr(__lowerCAmelCase , __lowerCAmelCase )[0] ) _UpperCAmelCase : Any = float(spearmanr(__lowerCAmelCase , __lowerCAmelCase )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : Any ) ->Optional[int]: '''simple docstring''' if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), "references": datasets.Value("int64" if self.config_name != "stsb" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : int ) ->str: '''simple docstring''' if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(lowerCamelCase__ , lowerCamelCase__ )} elif self.config_name == "stsb": return pearson_and_spearman(lowerCamelCase__ , lowerCamelCase__ ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(lowerCamelCase__ , lowerCamelCase__ ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(lowerCamelCase__ , lowerCamelCase__ )} else: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]" )
370
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = 'cuda' if torch.cuda.is_available() else 'cpu' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=100 , __lowerCAmelCase=" " ): _UpperCAmelCase : Any = text.split(__lowerCAmelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(__lowerCAmelCase ) , __lowerCAmelCase )] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase , _UpperCAmelCase : Dict = [], [] for title, text in zip(documents["title"] , documents["text"] ): if text is not None: for passage in split_text(__lowerCAmelCase ): titles.append(title if title is not None else "" ) texts.append(__lowerCAmelCase ) return {"title": titles, "text": texts} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : str = ctx_tokenizer( documents["title"] , documents["text"] , truncation=__lowerCAmelCase , padding="longest" , return_tensors="pt" )["input_ids"] _UpperCAmelCase : str = ctx_encoder(input_ids.to(device=__lowerCAmelCase ) , return_dict=__lowerCAmelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): ###################################### logger.info("Step 1 - Create the dataset" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way _UpperCAmelCase : Optional[int] = load_dataset( "csv" , data_files=[rag_example_args.csv_path] , split="train" , delimiter="\t" , column_names=["title", "text"] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words _UpperCAmelCase : Optional[int] = dataset.map(__lowerCAmelCase , batched=__lowerCAmelCase , num_proc=processing_args.num_proc ) # And compute the embeddings _UpperCAmelCase : Union[str, Any] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) _UpperCAmelCase : Dict = Features( {"text": Value("string" ), "title": Value("string" ), "embeddings": Sequence(Value("float32" ) )} ) # optional, save as float32 instead of float64 to save space _UpperCAmelCase : int = dataset.map( partial(__lowerCAmelCase , ctx_encoder=__lowerCAmelCase , ctx_tokenizer=__lowerCAmelCase ) , batched=__lowerCAmelCase , batch_size=processing_args.batch_size , features=__lowerCAmelCase , ) # And finally save your dataset _UpperCAmelCase : List[Any] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset" ) dataset.save_to_disk(__lowerCAmelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("Step 2 - Index the dataset" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search _UpperCAmelCase : Any = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("embeddings" , custom_index=__lowerCAmelCase ) # And save the index _UpperCAmelCase : List[str] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset_hnsw_index.faiss" ) dataset.get_index("embeddings" ).save(__lowerCAmelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" / "my_knowledge_dataset.csv" ) , metadata={"help": "Path to a tab-separated csv file with columns 'title' and 'text'"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Question that is passed as input to RAG. Default is 'What does Moses' rod turn into ?'."} , ) lowerCAmelCase : str = field( default="facebook/rag-sequence-nq" , metadata={"help": "The RAG model to use. Either 'facebook/rag-sequence-nq' or 'facebook/rag-token-nq'"} , ) lowerCAmelCase : str = field( default="facebook/dpr-ctx_encoder-multiset-base" , metadata={ "help": ( "The DPR context encoder model to use. Either 'facebook/dpr-ctx_encoder-single-nq-base' or" " 'facebook/dpr-ctx_encoder-multiset-base'" ) } , ) lowerCAmelCase : Optional[str] = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" ) , metadata={"help": "Path to a directory where the dataset passages and the index will be saved"} , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": "The number of processes to use to split the documents into passages. Default is single process." } , ) lowerCAmelCase : int = field( default=16 , metadata={ "help": "The batch size to use when computing the passages embeddings using the DPR context encoder." } , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : int = field( default=768 , metadata={"help": "The dimension of the embeddings to pass to the HNSW Faiss index."} , ) lowerCAmelCase : int = field( default=128 , metadata={ "help": ( "The number of bi-directional links created for every new element during the HNSW index construction." ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
322
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCamelCase__ = logging.get_logger(__name__) if is_vision_available(): import PIL class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : List[Any] = ["pixel_values"] def __init__( self : Optional[int] , lowerCamelCase__ : bool = True , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : PILImageResampling = PILImageResampling.BICUBIC , lowerCamelCase__ : bool = True , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : bool = True , lowerCamelCase__ : Union[int, float] = 1 / 2_55 , lowerCamelCase__ : bool = True , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , lowerCamelCase__ : bool = True , **lowerCamelCase__ : Dict , ) ->None: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = size if size is not None else {"shortest_edge": 2_24} _UpperCAmelCase : List[Any] = get_size_dict(lowerCamelCase__ , default_to_square=lowerCamelCase__ ) _UpperCAmelCase : List[Any] = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} _UpperCAmelCase : Union[str, Any] = get_size_dict(lowerCamelCase__ , default_to_square=lowerCamelCase__ , param_name="crop_size" ) _UpperCAmelCase : int = do_resize _UpperCAmelCase : Tuple = size _UpperCAmelCase : str = resample _UpperCAmelCase : List[str] = do_center_crop _UpperCAmelCase : int = crop_size _UpperCAmelCase : Dict = do_rescale _UpperCAmelCase : Any = rescale_factor _UpperCAmelCase : Optional[int] = do_normalize _UpperCAmelCase : Tuple = image_mean if image_mean is not None else OPENAI_CLIP_MEAN _UpperCAmelCase : str = image_std if image_std is not None else OPENAI_CLIP_STD _UpperCAmelCase : Tuple = do_convert_rgb def lowerCAmelCase__ ( self : str , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Dict[str, int] , lowerCamelCase__ : PILImageResampling = PILImageResampling.BICUBIC , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : List[Any] , ) ->np.ndarray: '''simple docstring''' _UpperCAmelCase : List[str] = get_size_dict(lowerCamelCase__ , default_to_square=lowerCamelCase__ ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) _UpperCAmelCase : List[str] = get_resize_output_image_size(lowerCamelCase__ , size=size["shortest_edge"] , default_to_square=lowerCamelCase__ ) return resize(lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Dict[str, int] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : str , ) ->np.ndarray: '''simple docstring''' _UpperCAmelCase : Optional[int] = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(lowerCamelCase__ , size=(size["height"], size["width"]) , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Union[int, float] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : int , ) ->Any: '''simple docstring''' return rescale(lowerCamelCase__ , scale=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Union[float, List[float]] , lowerCamelCase__ : Union[float, List[float]] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : Optional[Any] , ) ->np.ndarray: '''simple docstring''' return normalize(lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : ImageInput , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : PILImageResampling = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : int = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : float = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Optional[Union[str, TensorType]] = None , lowerCamelCase__ : Optional[ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase__ : List[str] , ) ->PIL.Image.Image: '''simple docstring''' _UpperCAmelCase : List[Any] = do_resize if do_resize is not None else self.do_resize _UpperCAmelCase : Optional[int] = size if size is not None else self.size _UpperCAmelCase : Optional[Any] = get_size_dict(lowerCamelCase__ , param_name="size" , default_to_square=lowerCamelCase__ ) _UpperCAmelCase : Tuple = resample if resample is not None else self.resample _UpperCAmelCase : Tuple = do_center_crop if do_center_crop is not None else self.do_center_crop _UpperCAmelCase : str = crop_size if crop_size is not None else self.crop_size _UpperCAmelCase : int = get_size_dict(lowerCamelCase__ , param_name="crop_size" , default_to_square=lowerCamelCase__ ) _UpperCAmelCase : Dict = do_rescale if do_rescale is not None else self.do_rescale _UpperCAmelCase : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCAmelCase : List[Any] = do_normalize if do_normalize is not None else self.do_normalize _UpperCAmelCase : Optional[Any] = image_mean if image_mean is not None else self.image_mean _UpperCAmelCase : Dict = image_std if image_std is not None else self.image_std _UpperCAmelCase : Optional[Any] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb _UpperCAmelCase : Tuple = make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: _UpperCAmelCase : Optional[int] = [convert_to_rgb(lowerCamelCase__ ) for image in images] # All transformations expect numpy arrays. _UpperCAmelCase : Tuple = [to_numpy_array(lowerCamelCase__ ) for image in images] if do_resize: _UpperCAmelCase : Tuple = [self.resize(image=lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ ) for image in images] if do_center_crop: _UpperCAmelCase : List[Any] = [self.center_crop(image=lowerCamelCase__ , size=lowerCamelCase__ ) for image in images] if do_rescale: _UpperCAmelCase : int = [self.rescale(image=lowerCamelCase__ , scale=lowerCamelCase__ ) for image in images] if do_normalize: _UpperCAmelCase : Any = [self.normalize(image=lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ ) for image in images] _UpperCAmelCase : str = [to_channel_dimension_format(lowerCamelCase__ , lowerCamelCase__ ) for image in images] _UpperCAmelCase : str = {"pixel_values": images} return BatchFeature(data=lowerCamelCase__ , tensor_type=lowerCamelCase__ )
371
'''simple docstring''' import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCamelCase__ = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class lowerCAmelCase__ ( unittest.TestCase ): @classmethod def lowerCAmelCase__ ( cls : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : Tuple = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def lowerCAmelCase__ ( cls : Union[str, Any] ) ->int: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def lowerCAmelCase__ ( self : int ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ , repo_id="test-config" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : Dict = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ , repo_id="valid_org/test-config-org" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : int = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' CustomConfig.register_for_auto_class() _UpperCAmelCase : int = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) _UpperCAmelCase : str = AutoConfig.from_pretrained(F"""{USER}/test-dynamic-config""" , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : List[str] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated _UpperCAmelCase : Any = c.n_embd + 1 # int _UpperCAmelCase : List[Any] = c.resid_pdrop + 1.0 # float _UpperCAmelCase : Tuple = not c.scale_attn_weights # bool _UpperCAmelCase : List[Any] = c.summary_type + "foo" # str c.update_from_string( F"""n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}""" ) self.assertEqual(lowerCamelCase__ , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(lowerCamelCase__ , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(lowerCamelCase__ , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(lowerCamelCase__ , c.summary_type , "mismatch for key: summary_type" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = PretrainedConfig() _UpperCAmelCase : Tuple = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase__ , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) _UpperCAmelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase__ , lowerCamelCase__ )] if len(lowerCamelCase__ ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" F""" {', '.join(lowerCamelCase__ )}.""" ) def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): # config is in subfolder, the following should not work without specifying the subfolder _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = mock.Mock() _UpperCAmelCase : List[str] = 5_00 _UpperCAmelCase : Dict = {} _UpperCAmelCase : Tuple = HTTPError _UpperCAmelCase : Any = {} # Download this model to make sure it's in the cache. _UpperCAmelCase : int = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=lowerCamelCase__ ) as mock_head: _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = AutoConfig.from_pretrained("bert-base-cased" ) _UpperCAmelCase : str = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 _UpperCAmelCase : Optional[int] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 _UpperCAmelCase : Dict = ["config.42.0.0.json"] _UpperCAmelCase : Union[str, Any] = 7_68 configuration.save_pretrained(lowerCamelCase__ ) shutil.move(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , os.path.join(lowerCamelCase__ , "config.42.0.0.json" ) ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = "hf-internal-testing/test-two-configs" import transformers as new_transformers _UpperCAmelCase : Any = "v4.0.0" _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers _UpperCAmelCase : List[Any] = "v3.0.0" _UpperCAmelCase : int = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(old_configuration.hidden_size , 7_68 )
322
0
'''simple docstring''' import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : int , lowerCamelCase__ : Any ) ->Union[str, Any]: '''simple docstring''' return F"""gaussian_noise_s={seed}_shape={'_'.join([str(lowerCamelCase__ ) for s in shape] )}.npy""" def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' super().tearDown() gc.collect() def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str=0 , lowerCamelCase__ : int=(4, 4, 64, 64) , lowerCamelCase__ : Union[str, Any]=False ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[int] = jnp.bfloataa if fpaa else jnp.floataa _UpperCAmelCase : Union[str, Any] = jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase__ , lowerCamelCase__ ) ) , dtype=lowerCamelCase__ ) return image def lowerCAmelCase__ ( self : int , lowerCamelCase__ : List[Any]=False , lowerCamelCase__ : Dict="CompVis/stable-diffusion-v1-4" ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = jnp.bfloataa if fpaa else jnp.floataa _UpperCAmelCase : Union[str, Any] = "bf16" if fpaa else None _UpperCAmelCase : Any = FlaxUNetaDConditionModel.from_pretrained( lowerCamelCase__ , subfolder="unet" , dtype=lowerCamelCase__ , revision=lowerCamelCase__ ) return model, params def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Optional[int]=0 , lowerCamelCase__ : List[Any]=(4, 77, 7_68) , lowerCamelCase__ : Optional[int]=False ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[str] = jnp.bfloataa if fpaa else jnp.floataa _UpperCAmelCase : Optional[Any] = jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase__ , lowerCamelCase__ ) ) , dtype=lowerCamelCase__ ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_3_2_3, -0.1_3_0_4, 0.0_8_1_3, -0.3_0_9_3, -0.0_9_1_9, -0.1_5_7_1, -0.1_1_2_5, -0.5_8_0_6]], [17, 0.5_5, [-0.0_8_3_1, -0.2_4_4_3, 0.0_9_0_1, -0.0_9_1_9, 0.3_3_9_6, 0.0_1_0_3, -0.3_7_4_3, 0.0_7_0_1]], [8, 0.8_9, [-0.4_8_6_3, 0.0_8_5_9, 0.0_8_7_5, -0.1_6_5_8, 0.9_1_9_9, -0.0_1_1_4, 0.4_8_3_9, 0.4_6_3_9]], [3, 10_00, [-0.5_6_4_9, 0.2_4_0_2, -0.5_5_1_8, 0.1_2_4_8, 1.1_3_2_8, -0.2_4_4_3, -0.0_3_2_5, -1.0_0_7_8]], # fmt: on ] ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[int] ) ->str: '''simple docstring''' _UpperCAmelCase : int = self.get_unet_model(model_id="CompVis/stable-diffusion-v1-4" , fpaa=lowerCamelCase__ ) _UpperCAmelCase : List[str] = self.get_latents(lowerCamelCase__ , fpaa=lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.get_encoder_hidden_states(lowerCamelCase__ , fpaa=lowerCamelCase__ ) _UpperCAmelCase : Tuple = model.apply( {"params": params} , lowerCamelCase__ , jnp.array(lowerCamelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=lowerCamelCase__ , ).sample assert sample.shape == latents.shape _UpperCAmelCase : Optional[Any] = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) _UpperCAmelCase : str = jnp.array(lowerCamelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_5_1_4, 0.0_8_0_7, 0.1_6_2_4, 0.1_0_1_6, -0.1_8_9_6, 0.0_2_6_3, 0.0_6_7_7, 0.2_3_1_0]], [17, 0.5_5, [0.1_1_6_4, -0.0_2_1_6, 0.0_1_7_0, 0.1_5_8_9, -0.3_1_2_0, 0.1_0_0_5, -0.0_5_8_1, -0.1_4_5_8]], [8, 0.8_9, [-0.1_7_5_8, -0.0_1_6_9, 0.1_0_0_4, -0.1_4_1_1, 0.1_3_1_2, 0.1_1_0_3, -0.1_9_9_6, 0.2_1_3_9]], [3, 10_00, [0.1_2_1_4, 0.0_3_5_2, -0.0_7_3_1, -0.1_5_6_2, -0.0_9_9_4, -0.0_9_0_6, -0.2_3_4_0, -0.0_5_3_9]], # fmt: on ] ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Dict , lowerCamelCase__ : Dict ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.get_unet_model(model_id="stabilityai/stable-diffusion-2" , fpaa=lowerCamelCase__ ) _UpperCAmelCase : Tuple = self.get_latents(lowerCamelCase__ , shape=(4, 4, 96, 96) , fpaa=lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.get_encoder_hidden_states(lowerCamelCase__ , shape=(4, 77, 10_24) , fpaa=lowerCamelCase__ ) _UpperCAmelCase : List[Any] = model.apply( {"params": params} , lowerCamelCase__ , jnp.array(lowerCamelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=lowerCamelCase__ , ).sample assert sample.shape == latents.shape _UpperCAmelCase : List[str] = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) _UpperCAmelCase : Dict = jnp.array(lowerCamelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1E-2 )
350
'''simple docstring''' from manim import * class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : List[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Optional[Any] = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Dict = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("CPU" , font_size=24 ) _UpperCAmelCase : Any = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(1 )] _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("GPU" , font_size=24 ) _UpperCAmelCase : str = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) gpu.align_to(lowerCamelCase__ , lowerCamelCase__ ) gpu.set_x(gpu.get_x() - 1 ) self.add(lowerCamelCase__ ) _UpperCAmelCase : List[str] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Any = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Model" , font_size=24 ) _UpperCAmelCase : Tuple = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.play( Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , ) _UpperCAmelCase : int = MarkupText( F"""First, an empty model skeleton is loaded\ninto <span fgcolor='{YELLOW}'>memory</span> without using much RAM.""" , font_size=24 , ) _UpperCAmelCase : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ , run_time=2.5 ) , Write(lowerCamelCase__ ) , Write(lowerCamelCase__ ) ) self.add(lowerCamelCase__ ) _UpperCAmelCase : int = [] _UpperCAmelCase : List[str] = [] _UpperCAmelCase : Dict = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : int = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ , opacity=0.7 ) cpu_target.move_to(lowerCamelCase__ ) cpu_target.generate_target() _UpperCAmelCase : Dict = 0.4_6 / 4 _UpperCAmelCase : Any = 0.4_6 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowerCamelCase__ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=lowerCamelCase__ , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=lowerCamelCase__ , buff=0.0 ) cpu_targs.append(lowerCamelCase__ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(lowerCamelCase__ ) ) second_animations.append(MoveToTarget(lowerCamelCase__ , run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(*lowerCamelCase__ ) self.wait()
322
0
'''simple docstring''' lowerCamelCase__ = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100_000)] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 100_000] number //= 100_000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowerCamelCase__ = [None] * 10_000_000 lowerCamelCase__ = True lowerCamelCase__ = False def __lowerCAmelCase (__lowerCAmelCase ): if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _UpperCAmelCase : str = chain(next_number(__lowerCAmelCase ) ) _UpperCAmelCase : Tuple = number_chain while number < 10_000_000: _UpperCAmelCase : Union[str, Any] = number_chain number *= 10 return number_chain def __lowerCAmelCase (__lowerCAmelCase = 10_000_000 ): for i in range(1 , __lowerCAmelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
351
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1_024 , __lowerCAmelCase=1_024 , __lowerCAmelCase=False , **__lowerCAmelCase ): _UpperCAmelCase : Any = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : List[str] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="train" , **__lowerCAmelCase ) _UpperCAmelCase : Dict = tok.pad_token_id def get_lens(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = tqdm( DataLoader(__lowerCAmelCase , batch_size=512 , num_workers=8 , shuffle=__lowerCAmelCase , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) _UpperCAmelCase : List[str] = [] for batch in dl: _UpperCAmelCase : Any = batch["input_ids"].ne(__lowerCAmelCase ).sum(1 ).tolist() _UpperCAmelCase : Tuple = batch["labels"].ne(__lowerCAmelCase ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__lowerCAmelCase , __lowerCAmelCase ): max_lens.append(max(__lowerCAmelCase , __lowerCAmelCase ) ) else: max_lens.extend(__lowerCAmelCase ) return max_lens _UpperCAmelCase : Dict = get_lens(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="val" , **__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = get_lens(__lowerCAmelCase ) pickle_save(__lowerCAmelCase , train_ds.len_file ) pickle_save(__lowerCAmelCase , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : List[str] = len(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = [[0] * n for i in range(__lowerCAmelCase )] for i in range(__lowerCAmelCase ): _UpperCAmelCase : List[Any] = y_points[i] for i in range(2 , __lowerCAmelCase ): for j in range(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Dict = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
352
'''simple docstring''' import pytest lowerCamelCase__ = '__dummy_dataset1__' lowerCamelCase__ = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = dataset_loading_script_name _UpperCAmelCase : Any = tmp_path / "datasets" / script_name script_dir.mkdir(parents=__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = script_dir / F"""{script_name}.py""" with open(__lowerCAmelCase , "w" ) as f: f.write(__lowerCAmelCase ) return str(__lowerCAmelCase )
322
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCamelCase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : List[str] = ["pixel_values"] def __init__( self : Tuple , lowerCamelCase__ : bool = True , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : PILImageResampling = PIL.Image.BICUBIC , lowerCamelCase__ : bool = True , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : Union[int, float] = 1 / 2_55 , lowerCamelCase__ : bool = True , lowerCamelCase__ : bool = True , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , **lowerCamelCase__ : Optional[int] , ) ->None: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : str = size if size is not None else {"height": 2_56, "width": 2_56} _UpperCAmelCase : int = get_size_dict(lowerCamelCase__ ) _UpperCAmelCase : Any = crop_size if crop_size is not None else {"height": 2_24, "width": 2_24} _UpperCAmelCase : Tuple = get_size_dict(lowerCamelCase__ , param_name="crop_size" ) _UpperCAmelCase : str = do_resize _UpperCAmelCase : List[str] = size _UpperCAmelCase : List[str] = resample _UpperCAmelCase : List[str] = do_center_crop _UpperCAmelCase : int = crop_size _UpperCAmelCase : int = do_rescale _UpperCAmelCase : Tuple = rescale_factor _UpperCAmelCase : Tuple = do_normalize _UpperCAmelCase : str = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _UpperCAmelCase : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Dict[str, int] , lowerCamelCase__ : PILImageResampling = PIL.Image.BICUBIC , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : List[str] , ) ->np.ndarray: '''simple docstring''' _UpperCAmelCase : Optional[int] = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return resize( lowerCamelCase__ , size=(size["height"], size["width"]) , resample=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Dict[str, int] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : Any , ) ->np.ndarray: '''simple docstring''' _UpperCAmelCase : Optional[int] = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(lowerCamelCase__ , size=(size["height"], size["width"]) , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Union[int, float] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : Tuple , ) ->str: '''simple docstring''' return rescale(lowerCamelCase__ , scale=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : Union[float, List[float]] , lowerCamelCase__ : Union[float, List[float]] , lowerCamelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase__ : List[Any] , ) ->np.ndarray: '''simple docstring''' return normalize(lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : ImageInput , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : Tuple=None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict[str, int] = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : float = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , lowerCamelCase__ : Optional[Union[float, List[float]]] = None , lowerCamelCase__ : Optional[Union[str, TensorType]] = None , lowerCamelCase__ : ChannelDimension = ChannelDimension.FIRST , **lowerCamelCase__ : Dict , ) ->PIL.Image.Image: '''simple docstring''' _UpperCAmelCase : Optional[int] = do_resize if do_resize is not None else self.do_resize _UpperCAmelCase : List[Any] = resample if resample is not None else self.resample _UpperCAmelCase : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop _UpperCAmelCase : List[str] = do_rescale if do_rescale is not None else self.do_rescale _UpperCAmelCase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCAmelCase : List[str] = do_normalize if do_normalize is not None else self.do_normalize _UpperCAmelCase : Union[str, Any] = image_mean if image_mean is not None else self.image_mean _UpperCAmelCase : Any = image_std if image_std is not None else self.image_std _UpperCAmelCase : Union[str, Any] = size if size is not None else self.size _UpperCAmelCase : List[str] = get_size_dict(lowerCamelCase__ ) _UpperCAmelCase : str = crop_size if crop_size is not None else self.crop_size _UpperCAmelCase : Optional[Any] = get_size_dict(lowerCamelCase__ , param_name="crop_size" ) _UpperCAmelCase : List[Any] = make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _UpperCAmelCase : List[str] = [to_numpy_array(lowerCamelCase__ ) for image in images] if do_resize: _UpperCAmelCase : Union[str, Any] = [self.resize(image=lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ ) for image in images] if do_center_crop: _UpperCAmelCase : Optional[int] = [self.center_crop(image=lowerCamelCase__ , size=lowerCamelCase__ ) for image in images] if do_rescale: _UpperCAmelCase : Optional[int] = [self.rescale(image=lowerCamelCase__ , scale=lowerCamelCase__ ) for image in images] if do_normalize: _UpperCAmelCase : str = [self.normalize(image=lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ ) for image in images] _UpperCAmelCase : Dict = [to_channel_dimension_format(lowerCamelCase__ , lowerCamelCase__ ) for image in images] _UpperCAmelCase : Optional[int] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase__ , tensor_type=lowerCamelCase__ )
353
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase__ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCamelCase__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCamelCase__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCamelCase__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"] , reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ] , ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] ) ->int: '''simple docstring''' import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : int=0.9 , lowerCamelCase__ : Dict=3 , lowerCamelCase__ : Dict=0.5 ) ->Any: '''simple docstring''' if NLTK_VERSION >= version.Version("3.6.5" ): _UpperCAmelCase : Dict = [ meteor_score.single_meteor_score( word_tokenize(lowerCamelCase__ ) , word_tokenize(lowerCamelCase__ ) , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] else: _UpperCAmelCase : Optional[int] = [ meteor_score.single_meteor_score(lowerCamelCase__ , lowerCamelCase__ , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] return {"meteor": np.mean(lowerCamelCase__ )}
322
0
'''simple docstring''' from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( 'pipelines_utils', '0.22.0', 'Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.', standard_warn=False, stacklevel=3, )
354
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline lowerCamelCase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : int ) ->str: '''simple docstring''' if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Union[str, Any] = [label.strip() for label in labels.split("," ) if label.strip()] return labels def __call__( self : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] ) ->str: '''simple docstring''' if len(lowerCamelCase__ ) == 0 or len(lowerCamelCase__ ) == 0: raise ValueError("You must include at least one label and at least one sequence." ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( "The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. " "Make sure the passed template includes formatting syntax such as {{}} where the label should go." ).format(lowerCamelCase__ ) ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Optional[Any] = [sequences] _UpperCAmelCase : int = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(lowerCamelCase__ )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(UpperCAmelCase__ ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Union[str, Any] , lowerCamelCase__ : Optional[Any]=ZeroShotClassificationArgumentHandler() , *lowerCamelCase__ : List[str] , **lowerCamelCase__ : Any ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = args_parser super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) if self.entailment_id == -1: logger.warning( "Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to " "-1. Define a descriptive label2id mapping in the model config to ensure correct outputs." ) @property def lowerCAmelCase__ ( self : Any ) ->Union[str, Any]: '''simple docstring''' for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("entail" ): return ind return -1 def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Tuple , lowerCamelCase__ : int=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : str=TruncationStrategy.ONLY_FIRST , **lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : int = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( "Tokenizer was not supporting padding necessary for zero-shot, attempting to use " " `pad_token=eos_token`" ) _UpperCAmelCase : Optional[Any] = self.tokenizer.eos_token try: _UpperCAmelCase : List[str] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , ) except Exception as e: if "too short" in str(lowerCamelCase__ ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. _UpperCAmelCase : List[Any] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def lowerCAmelCase__ ( self : int , **lowerCamelCase__ : Union[str, Any] ) ->Tuple: '''simple docstring''' if kwargs.get("multi_class" , lowerCamelCase__ ) is not None: _UpperCAmelCase : int = kwargs["multi_class"] logger.warning( "The `multi_class` argument has been deprecated and renamed to `multi_label`. " "`multi_class` will be removed in a future version of Transformers." ) _UpperCAmelCase : Dict = {} if "candidate_labels" in kwargs: _UpperCAmelCase : List[Any] = self._args_parser._parse_labels(kwargs["candidate_labels"] ) if "hypothesis_template" in kwargs: _UpperCAmelCase : Dict = kwargs["hypothesis_template"] _UpperCAmelCase : List[str] = {} if "multi_label" in kwargs: _UpperCAmelCase : Optional[Any] = kwargs["multi_label"] return preprocess_params, {}, postprocess_params def __call__( self : int , lowerCamelCase__ : Union[str, List[str]] , *lowerCamelCase__ : str , **lowerCamelCase__ : Optional[Any] , ) ->Optional[int]: '''simple docstring''' if len(lowerCamelCase__ ) == 0: pass elif len(lowerCamelCase__ ) == 1 and "candidate_labels" not in kwargs: _UpperCAmelCase : int = args[0] else: raise ValueError(F"""Unable to understand extra arguments {args}""" ) return super().__call__(lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Any=None , lowerCamelCase__ : str="This example is {}." ) ->Tuple: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Optional[int] = self._args_parser(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for i, (candidate_label, sequence_pair) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): _UpperCAmelCase : Optional[int] = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(lowerCamelCase__ ) - 1, **model_input, } def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Optional[int] ) ->int: '''simple docstring''' _UpperCAmelCase : Dict = inputs["candidate_label"] _UpperCAmelCase : Optional[int] = inputs["sequence"] _UpperCAmelCase : Dict = {k: inputs[k] for k in self.tokenizer.model_input_names} _UpperCAmelCase : List[Any] = self.model(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = { "candidate_label": candidate_label, "sequence": sequence, "is_last": inputs["is_last"], **outputs, } return model_outputs def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple=False ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = [outputs["candidate_label"] for outputs in model_outputs] _UpperCAmelCase : Any = [outputs["sequence"] for outputs in model_outputs] _UpperCAmelCase : Optional[int] = np.concatenate([output["logits"].numpy() for output in model_outputs] ) _UpperCAmelCase : Optional[Any] = logits.shape[0] _UpperCAmelCase : Any = len(lowerCamelCase__ ) _UpperCAmelCase : str = N // n _UpperCAmelCase : str = logits.reshape((num_sequences, n, -1) ) if multi_label or len(lowerCamelCase__ ) == 1: # softmax over the entailment vs. contradiction dim for each label independently _UpperCAmelCase : int = self.entailment_id _UpperCAmelCase : List[Any] = -1 if entailment_id == 0 else 0 _UpperCAmelCase : str = reshaped_outputs[..., [contradiction_id, entailment_id]] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : str = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels _UpperCAmelCase : int = reshaped_outputs[..., self.entailment_id] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
322
0
'''simple docstring''' from manim import * class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Union[str, Any] ) ->int: '''simple docstring''' _UpperCAmelCase : Optional[Any] = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Union[str, Any] = Rectangle(height=0.2_5 , width=0.2_5 ) _UpperCAmelCase : str = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _UpperCAmelCase : List[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Any = [mem.copy() for i in range(6 )] _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Tuple = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = Text("CPU" , font_size=24 ) _UpperCAmelCase : List[Any] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Tuple = [mem.copy() for i in range(4 )] _UpperCAmelCase : Optional[int] = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = Text("GPU" , font_size=24 ) _UpperCAmelCase : List[str] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) gpu.move_to([-1, -1, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : str = [mem.copy() for i in range(6 )] _UpperCAmelCase : Optional[int] = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = Text("Model" , font_size=24 ) _UpperCAmelCase : Optional[Any] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = [] _UpperCAmelCase : Union[str, Any] = [] _UpperCAmelCase : str = [] for i, rect in enumerate(lowerCamelCase__ ): rect.set_stroke(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = Rectangle(height=0.4_6 / 4 , width=0.4_6 / 3 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowerCamelCase__ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=lowerCamelCase__ , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=lowerCamelCase__ , buff=0.0 ) self.add(lowerCamelCase__ ) model_cpu_arr.append(lowerCamelCase__ ) self.add(*lowerCamelCase__ , *lowerCamelCase__ , *lowerCamelCase__ ) _UpperCAmelCase : List[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : int = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Loaded Checkpoint" , font_size=24 ) _UpperCAmelCase : Optional[int] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) checkpoint.move_to([3, 0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Tuple = [] _UpperCAmelCase : Optional[Any] = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : Dict = fill.copy().set_fill(lowerCamelCase__ , opacity=0.7 ) target.move_to(lowerCamelCase__ ) ckpt_arr.append(lowerCamelCase__ ) _UpperCAmelCase : int = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(lowerCamelCase__ ) self.add(*lowerCamelCase__ , *lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Any = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(lowerCamelCase__ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Any = MarkupText( F"""Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) _UpperCAmelCase : str = [meta_mem.copy() for i in range(6 )] _UpperCAmelCase : Union[str, Any] = [meta_mem.copy() for i in range(6 )] _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Any = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Union[str, Any] = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : List[str] = Text("Disk" , font_size=24 ) _UpperCAmelCase : Union[str, Any] = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) disk.move_to([-4.0, -1.2_5, 0] ) self.play(Write(lowerCamelCase__ , run_time=3 ) , Write(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) ) _UpperCAmelCase : List[str] = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : Dict = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(lowerCamelCase__ , run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(FadeOut(lowerCamelCase__ ) ) _UpperCAmelCase : Any = MarkupText(F"""Then, the checkpoint is removed from memory\nthrough garbage collection.""" , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ , run_time=3 ) ) self.play( FadeOut(lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ , *lowerCamelCase__ ) , ) self.wait()
355
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase = 4_000_000 ): _UpperCAmelCase : List[Any] = [] _UpperCAmelCase , _UpperCAmelCase : Dict = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__lowerCAmelCase ) _UpperCAmelCase , _UpperCAmelCase : Any = b, a + b return sum(__lowerCAmelCase ) if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' import collections import os import re from pathlib import Path lowerCamelCase__ : List[Any] = 'src/transformers' # Matches is_xxx_available() lowerCamelCase__ : List[str] = re.compile(r'is\_([a-z_]*)_available()') # Catches a one-line _import_struct = {xxx} lowerCamelCase__ : Optional[Any] = re.compile(r'^_import_structure\s+=\s+\{([^\}]+)\}') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] lowerCamelCase__ : Any = re.compile(r'\s+"\S*":\s+\[([^\]]*)\]') # Catches a line if not is_foo_available lowerCamelCase__ : Any = re.compile(r'^\s*if\s+not\s+is\_[a-z_]*\_available\(\)') # Catches a line _import_struct["bla"].append("foo") lowerCamelCase__ : Tuple = re.compile(r'^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] lowerCamelCase__ : Dict = re.compile(r'^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]') # Catches a line with an object between quotes and a comma: "MyModel", lowerCamelCase__ : Optional[int] = re.compile(r'^\s+"([^"]+)",') # Catches a line with objects between brackets only: ["foo", "bar"], lowerCamelCase__ : List[Any] = re.compile(r'^\s+\[([^\]]+)\]') # Catches a line with from foo import bar, bla, boo lowerCamelCase__ : List[Any] = re.compile(r'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') # Catches a line with try: lowerCamelCase__ : str = re.compile(r'^\s*try:') # Catches a line with else: lowerCamelCase__ : Dict = re.compile(r'^\s*else:') def __lowerCAmelCase (__lowerCAmelCase ): if _re_test_backend.search(__lowerCAmelCase ) is None: return None _UpperCAmelCase : Optional[int] = [b[0] for b in _re_backend.findall(__lowerCAmelCase )] backends.sort() return "_and_".join(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "r" , encoding="utf-8" , newline="\n" ) as f: _UpperCAmelCase : str = f.readlines() _UpperCAmelCase : Union[str, Any] = 0 while line_index < len(__lowerCAmelCase ) and not lines[line_index].startswith("_import_structure = {" ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(__lowerCAmelCase ): return None # First grab the objects without a specific backend in _import_structure _UpperCAmelCase : Dict = [] while not lines[line_index].startswith("if TYPE_CHECKING" ) and find_backend(lines[line_index] ) is None: _UpperCAmelCase : int = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(__lowerCAmelCase ): _UpperCAmelCase : Tuple = _re_one_line_import_struct.search(__lowerCAmelCase ).groups()[0] _UpperCAmelCase : Optional[int] = re.findall(R"\[([^\]]+)\]" , __lowerCAmelCase ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(", " )] ) line_index += 1 continue _UpperCAmelCase : int = _re_import_struct_key_value.search(__lowerCAmelCase ) if single_line_import_search is not None: _UpperCAmelCase : str = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(", " ) if len(__lowerCAmelCase ) > 0] objects.extend(__lowerCAmelCase ) elif line.startswith(" " * 8 + "\"" ): objects.append(line[9:-3] ) line_index += 1 _UpperCAmelCase : List[Any] = {"none": objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith("if TYPE_CHECKING" ): # If the line is an if not is_backend_available, we grab all objects associated. _UpperCAmelCase : str = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _UpperCAmelCase : List[str] = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _UpperCAmelCase : Any = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(" " * 4 ): _UpperCAmelCase : Dict = lines[line_index] if _re_import_struct_add_one.search(__lowerCAmelCase ) is not None: objects.append(_re_import_struct_add_one.search(__lowerCAmelCase ).groups()[0] ) elif _re_import_struct_add_many.search(__lowerCAmelCase ) is not None: _UpperCAmelCase : List[Any] = _re_import_struct_add_many.search(__lowerCAmelCase ).groups()[0].split(", " ) _UpperCAmelCase : Optional[int] = [obj[1:-1] for obj in imports if len(__lowerCAmelCase ) > 0] objects.extend(__lowerCAmelCase ) elif _re_between_brackets.search(__lowerCAmelCase ) is not None: _UpperCAmelCase : Any = _re_between_brackets.search(__lowerCAmelCase ).groups()[0].split(", " ) _UpperCAmelCase : Optional[int] = [obj[1:-1] for obj in imports if len(__lowerCAmelCase ) > 0] objects.extend(__lowerCAmelCase ) elif _re_quote_object.search(__lowerCAmelCase ) is not None: objects.append(_re_quote_object.search(__lowerCAmelCase ).groups()[0] ) elif line.startswith(" " * 8 + "\"" ): objects.append(line[9:-3] ) elif line.startswith(" " * 12 + "\"" ): objects.append(line[13:-3] ) line_index += 1 _UpperCAmelCase : int = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend _UpperCAmelCase : Any = [] while ( line_index < len(__lowerCAmelCase ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith("else" ) ): _UpperCAmelCase : Optional[int] = lines[line_index] _UpperCAmelCase : Union[str, Any] = _re_import.search(__lowerCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 8 ): objects.append(line[8:-2] ) line_index += 1 _UpperCAmelCase : Optional[Any] = {"none": objects} # Let's continue with backend-specific objects while line_index < len(__lowerCAmelCase ): # If the line is an if is_backend_available, we grab all objects associated. _UpperCAmelCase : str = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _UpperCAmelCase : Any = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _UpperCAmelCase : Any = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(" " * 8 ): _UpperCAmelCase : List[str] = lines[line_index] _UpperCAmelCase : Dict = _re_import.search(__lowerCAmelCase ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 12 ): objects.append(line[12:-2] ) line_index += 1 _UpperCAmelCase : str = objects else: line_index += 1 return import_dict_objects, type_hint_objects def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): def find_duplicates(__lowerCAmelCase ): return [k for k, v in collections.Counter(__lowerCAmelCase ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] _UpperCAmelCase : Dict = [] for key in import_dict_objects.keys(): _UpperCAmelCase : Optional[int] = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F"""Duplicate _import_structure definitions for: {duplicate_imports}""" ) _UpperCAmelCase : Any = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F"""Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}""" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): _UpperCAmelCase : Optional[int] = "base imports" if key == "none" else F"""{key} backend""" errors.append(F"""Differences for {name}:""" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F""" {a} in TYPE_HINT but not in _import_structure.""" ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F""" {a} in _import_structure but not in TYPE_HINT.""" ) return errors def __lowerCAmelCase (): _UpperCAmelCase : int = [] for root, _, files in os.walk(__lowerCAmelCase ): if "__init__.py" in files: _UpperCAmelCase : int = os.path.join(__lowerCAmelCase , "__init__.py" ) _UpperCAmelCase : List[Any] = parse_init(__lowerCAmelCase ) if objects is not None: _UpperCAmelCase : Any = analyze_results(*__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: _UpperCAmelCase : Optional[Any] = F"""Problem in {fname}, both halves do not define the same objects.\n{errors[0]}""" failures.append("\n".join(__lowerCAmelCase ) ) if len(__lowerCAmelCase ) > 0: raise ValueError("\n\n".join(__lowerCAmelCase ) ) def __lowerCAmelCase (): _UpperCAmelCase : Dict = [] for path, directories, files in os.walk(__lowerCAmelCase ): for folder in directories: # Ignore private modules if folder.startswith("_" ): directories.remove(__lowerCAmelCase ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(__lowerCAmelCase ) / folder).glob("*.py" ) ) ) == 0: continue _UpperCAmelCase : Dict = str((Path(__lowerCAmelCase ) / folder).relative_to(__lowerCAmelCase ) ) _UpperCAmelCase : Union[str, Any] = short_path.replace(os.path.sep , "." ) submodules.append(__lowerCAmelCase ) for fname in files: if fname == "__init__.py": continue _UpperCAmelCase : Dict = str((Path(__lowerCAmelCase ) / fname).relative_to(__lowerCAmelCase ) ) _UpperCAmelCase : List[Any] = short_path.replace(".py" , "" ).replace(os.path.sep , "." ) if len(submodule.split("." ) ) == 1: submodules.append(__lowerCAmelCase ) return submodules lowerCamelCase__ : Union[str, Any] = [ 'convert_pytorch_checkpoint_to_tf2', 'modeling_flax_pytorch_utils', 'models.esm.openfold_utils', ] def __lowerCAmelCase (): # This is to make sure the transformers module imported is the one in the repo. from transformers.utils import direct_transformers_import _UpperCAmelCase : Any = direct_transformers_import(__lowerCAmelCase ) _UpperCAmelCase : Any = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(__lowerCAmelCase , "__init__.py" ) , "r" ) as f: _UpperCAmelCase : Tuple = f.read() import_structure_keys.update(set(re.findall(R"import_structure\[\"([^\"]*)\"\]" , __lowerCAmelCase ) ) ) _UpperCAmelCase : Union[str, Any] = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(__lowerCAmelCase ) > 0: _UpperCAmelCase : List[Any] = "\n".join(F"""- {module}""" for module in module_not_registered ) raise ValueError( "The following submodules are not properly registed in the main init of Transformers:\n" F"""{list_of_modules}\n""" "Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value." ) if __name__ == "__main__": check_all_inits() check_submodules()
356
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[str]=13 , lowerCamelCase__ : Optional[Any]=7 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : int=99 , lowerCamelCase__ : int=32 , lowerCamelCase__ : List[str]=5 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : Optional[int]=37 , lowerCamelCase__ : Tuple="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Union[str, Any]=0.1 , lowerCamelCase__ : Optional[int]=5_12 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : str=2 , lowerCamelCase__ : Union[str, Any]=0.0_2 , lowerCamelCase__ : Tuple=4 , ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = parent _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Optional[int] = seq_length _UpperCAmelCase : int = is_training _UpperCAmelCase : Dict = use_attention_mask _UpperCAmelCase : Optional[Any] = use_token_type_ids _UpperCAmelCase : int = use_labels _UpperCAmelCase : Optional[int] = vocab_size _UpperCAmelCase : Any = hidden_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : List[Any] = num_attention_heads _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : int = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob _UpperCAmelCase : Union[str, Any] = max_position_embeddings _UpperCAmelCase : Tuple = type_vocab_size _UpperCAmelCase : List[Any] = type_sequence_label_size _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Dict = num_choices def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Dict = None if self.use_attention_mask: _UpperCAmelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCAmelCase : Union[str, Any] = None if self.use_token_type_ids: _UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCAmelCase : int = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = config_and_inputs _UpperCAmelCase : str = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Optional[int] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase__ ( self : Optional[int] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = FlaxAlbertModelTester(self ) @slow def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCAmelCase : List[str] = model_class_name.from_pretrained("albert-base-v2" ) _UpperCAmelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = FlaxAlbertModel.from_pretrained("albert-base-v2" ) _UpperCAmelCase : List[Any] = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) _UpperCAmelCase : Optional[int] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCAmelCase : Dict = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )[0] _UpperCAmelCase : List[Any] = (1, 11, 7_68) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : str = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase__ , atol=1E-4 ) )
322
0
'''simple docstring''' import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : int = CLIPTokenizer lowerCAmelCase : Dict = CLIPTokenizerFast lowerCAmelCase : Any = True lowerCAmelCase : Any = {} lowerCAmelCase : List[Any] = False def lowerCAmelCase__ ( self : Any ) ->Optional[int]: '''simple docstring''' super().setUp() # fmt: off _UpperCAmelCase : Optional[Any] = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on _UpperCAmelCase : Optional[Any] = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _UpperCAmelCase : Tuple = ["#version: 0.2", "l o", "lo w</w>", "e r</w>"] _UpperCAmelCase : List[str] = {"unk_token": "<unk>"} _UpperCAmelCase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) _UpperCAmelCase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] , **lowerCamelCase__ : Dict ) ->Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Any , **lowerCamelCase__ : Tuple ) ->Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Optional[int] ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = "lower newer" _UpperCAmelCase : int = "lower newer" return input_text, output_text def lowerCAmelCase__ ( self : Tuple ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _UpperCAmelCase : str = "lower newer" _UpperCAmelCase : Optional[int] = ["lo", "w", "er</w>", "n", "e", "w", "er</w>"] _UpperCAmelCase : Union[str, Any] = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = tokens + [tokenizer.unk_token] _UpperCAmelCase : Tuple = [10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ ) @require_ftfy def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _UpperCAmelCase : Union[str, Any] = self.tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = "A\n'll 11p223RF☆ho!!to?'d'd''d of a cat to-$''d." _UpperCAmelCase : Tuple = tokenizer_s.tokenize(lowerCamelCase__ ) _UpperCAmelCase : Dict = tokenizer_r.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways _UpperCAmelCase : List[Any] = "xa\u0303y" + " " + "x\xe3y" _UpperCAmelCase : Optional[Any] = tokenizer_s.tokenize(lowerCamelCase__ ) _UpperCAmelCase : List[str] = tokenizer_r.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Test that the tokenization is identical on unicode of space type _UpperCAmelCase : int = [ "\u0009", # (horizontal tab, '\t') "\u000B", # (vertical tab) "\u000C", # (form feed) "\u0020", # (space, ' ') "\u200E", # (left-to-right mark):w "\u200F", # (right-to-left mark) ] for unicode_seq in spaces_unicodes: _UpperCAmelCase : List[str] = tokenizer_s.tokenize(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = tokenizer_r.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Test that the tokenization is identical on unicode of line break type _UpperCAmelCase : int = [ "\u000A", # (line feed, '\n') "\r\n", # (carriage return and line feed, '\r\n') "\u000D", # (carriage return, '\r') "\r", # (carriage return, '\r') "\u000D", # (carriage return, '\r') "\u2028", # (line separator) "\u2029", # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: _UpperCAmelCase : Tuple = tokenizer_s.tokenize(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = tokenizer_r.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Any ) ->List[Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _UpperCAmelCase : int = "hello" # `hello` is a token in the vocabulary of `pretrained_name` _UpperCAmelCase : List[Any] = F"""{text_of_1_token} {text_of_1_token}""" _UpperCAmelCase : Tuple = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__ , use_fast=lowerCamelCase__ , ) _UpperCAmelCase : str = tokenizer_r(lowerCamelCase__ , return_offsets_mapping=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowerCamelCase__ ) + 1, len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )) , ) _UpperCAmelCase : Optional[int] = F""" {text}""" _UpperCAmelCase : Any = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__ , use_fast=lowerCamelCase__ , ) _UpperCAmelCase : Tuple = tokenizer_r(lowerCamelCase__ , return_offsets_mapping=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowerCamelCase__ ) + 1, 1 + len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )) , ) def lowerCAmelCase__ ( self : List[Any] ) ->int: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ) as context: self.rust_tokenizer_class.from_pretrained("robot-test/old-clip-tokenizer" ) self.assertTrue( context.exception.args[0].startswith( "The `backend_tokenizer` provided does not match the expected format." ) ) @require_ftfy def lowerCAmelCase__ ( self : Union[str, Any] ) ->Dict: '''simple docstring''' super().test_tokenization_python_rust_equals() def lowerCAmelCase__ ( self : str ) ->Dict: '''simple docstring''' pass
357
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-classification/requirements.txt') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "rb" ) as f: _UpperCAmelCase : List[str] = Image.open(__lowerCAmelCase ) return im.convert("RGB" ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={ "help": "Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub)." } , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the training data."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the validation data."} ) lowerCAmelCase : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default="google/vit-base-patch16-224-in21k" , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCAmelCase__ )} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) lowerCAmelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCAmelCase : str = field(default=UpperCAmelCase__ , metadata={"help": "Name or path of preprocessor config."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = torch.stack([example["pixel_values"] for example in examples] ) _UpperCAmelCase : Tuple = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def __lowerCAmelCase (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , __lowerCAmelCase , __lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(__lowerCAmelCase ) transformers.utils.logging.set_verbosity(__lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _UpperCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCAmelCase : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCAmelCase : List[Any] = {} if data_args.train_dir is not None: _UpperCAmelCase : str = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCAmelCase : Optional[Any] = os.path.join(data_args.validation_dir , "**" ) _UpperCAmelCase : Any = load_dataset( "imagefolder" , data_files=__lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCAmelCase : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __lowerCAmelCase ) and data_args.train_val_split > 0.0: _UpperCAmelCase : List[Any] = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCAmelCase : List[str] = split["train"] _UpperCAmelCase : Union[str, Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase : Optional[int] = dataset["train"].features["labels"].names _UpperCAmelCase , _UpperCAmelCase : int = {}, {} for i, label in enumerate(__lowerCAmelCase ): _UpperCAmelCase : int = str(__lowerCAmelCase ) _UpperCAmelCase : str = label # Load the accuracy metric from the datasets package _UpperCAmelCase : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(__lowerCAmelCase ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowerCAmelCase ) , labelaid=__lowerCAmelCase , idalabel=__lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCAmelCase : Dict = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCAmelCase : int = image_processor.size["shortest_edge"] else: _UpperCAmelCase : int = (image_processor.size["height"], image_processor.size["width"]) _UpperCAmelCase : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCAmelCase : Optional[int] = Compose( [ RandomResizedCrop(__lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCAmelCase : Union[str, Any] = Compose( [ Resize(__lowerCAmelCase ), CenterCrop(__lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCAmelCase : Dict = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(__lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCAmelCase : Optional[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(__lowerCAmelCase ) # Initalize our trainer _UpperCAmelCase : Union[str, Any] = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=__lowerCAmelCase , tokenizer=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : Any = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : int = last_checkpoint _UpperCAmelCase : Dict = trainer.train(resume_from_checkpoint=__lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowerCAmelCase ) trainer.save_metrics("eval" , __lowerCAmelCase ) # Write model card and (optionally) push to hub _UpperCAmelCase : int = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowerCAmelCase ) else: trainer.create_model_card(**__lowerCAmelCase ) if __name__ == "__main__": main()
322
0
def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[Any] = len(__lowerCAmelCase ) _UpperCAmelCase : Tuple = sum(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _UpperCAmelCase : Any = True for i in range(1 , s + 1 ): _UpperCAmelCase : List[Any] = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): _UpperCAmelCase : Optional[int] = dp[i][j - 1] if arr[i - 1] <= j: _UpperCAmelCase : Any = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: _UpperCAmelCase : List[Any] = s - 2 * j break return diff
358
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase__ = logging.get_logger(__name__) # General docstring lowerCamelCase__ = 'RegNetConfig' # Base docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = [1, 1_088, 7, 7] # Image classification docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = 'tabby, tabby cat' lowerCamelCase__ = [ 'facebook/regnet-y-040', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 3 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[str] = "relu" , **lowerCamelCase__ : Tuple , ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _UpperCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _UpperCAmelCase : Dict = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=lowerCamelCase__ , strides=lowerCamelCase__ , padding="VALID" , groups=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" , ) _UpperCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) _UpperCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.convolution(self.padding(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = self.normalization(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = config.num_channels _UpperCAmelCase : Any = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = shape_list(lowerCamelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _UpperCAmelCase : Optional[Any] = tf.transpose(lowerCamelCase__ , perm=(0, 2, 3, 1) ) _UpperCAmelCase : List[Any] = self.embedder(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : int = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=1 , strides=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" ) _UpperCAmelCase : Any = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False ) ->tf.Tensor: '''simple docstring''' return self.normalization(self.convolution(lowerCamelCase__ ) , training=lowerCamelCase__ ) class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : int , **lowerCamelCase__ : Optional[int] ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) _UpperCAmelCase : int = [ tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.pooler(lowerCamelCase__ ) for layer_module in self.attention: _UpperCAmelCase : str = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : Any ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = in_channels != out_channels or stride != 1 _UpperCAmelCase : List[str] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : List[str] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _UpperCAmelCase : Optional[int] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.2" ), ] _UpperCAmelCase : Union[str, Any] = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = hidden_state for layer_module in self.layers: _UpperCAmelCase : List[Any] = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : List[Any] , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Optional[int] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : Union[str, Any] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) _UpperCAmelCase : List[Any] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(lowerCamelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.3" ), ] _UpperCAmelCase : int = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str ) ->Any: '''simple docstring''' _UpperCAmelCase : int = hidden_state for layer_module in self.layers: _UpperCAmelCase : Tuple = layer_module(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : Tuple = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : str = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer _UpperCAmelCase : List[str] = [ # downsampling is done in the first layer with stride of 2 layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , name="layers.0" ), *[layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , name=F"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] ) ->List[str]: '''simple docstring''' for layer_module in self.layers: _UpperCAmelCase : Optional[int] = layer_module(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : int ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) _UpperCAmelCase : Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCamelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , depth=lowerCamelCase__ , name=F"""stages.{i+1}""" ) ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = True ) ->TFBaseModelOutputWithNoAttention: '''simple docstring''' _UpperCAmelCase : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase : Optional[Any] = hidden_states + (hidden_state,) _UpperCAmelCase : Dict = stage_module(lowerCamelCase__ ) if output_hidden_states: _UpperCAmelCase : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase__ , hidden_states=lowerCamelCase__ ) @keras_serializable class lowerCAmelCase__ ( tf.keras.layers.Layer ): lowerCAmelCase : Optional[Any] = RegNetConfig def __init__( self : Union[str, Any] , lowerCamelCase__ : Any , **lowerCamelCase__ : str ) ->int: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = config _UpperCAmelCase : Union[str, Any] = TFRegNetEmbeddings(lowerCamelCase__ , name="embedder" ) _UpperCAmelCase : Union[str, Any] = TFRegNetEncoder(lowerCamelCase__ , name="encoder" ) _UpperCAmelCase : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) @unpack_inputs def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : bool = False , ) ->TFBaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' _UpperCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.embedder(lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : str = self.encoder( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : Dict = encoder_outputs[0] _UpperCAmelCase : Dict = self.pooler(lowerCamelCase__ ) # Change to NCHW output format have uniformity in the modules _UpperCAmelCase : Union[str, Any] = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) _UpperCAmelCase : Tuple = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _UpperCAmelCase : List[str] = tuple([tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase__ , pooler_output=lowerCamelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = RegNetConfig lowerCAmelCase : Tuple = "regnet" lowerCAmelCase : Union[str, Any] = "pixel_values" @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} lowerCamelCase__ = r'\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCamelCase__ = r'\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : Any , **lowerCamelCase__ : List[str] ) ->Optional[int]: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Any=False , ) ->Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( pixel_values=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : List[Any] , **lowerCamelCase__ : Union[str, Any] ) ->Any: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = config.num_labels _UpperCAmelCase : Dict = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) # classification head _UpperCAmelCase : str = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict=False , ) ->Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : str = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : int = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase : Dict = self.classifier[0](lowerCamelCase__ ) _UpperCAmelCase : str = self.classifier[1](lowerCamelCase__ ) _UpperCAmelCase : Tuple = None if labels is None else self.hf_compute_loss(labels=lowerCamelCase__ , logits=lowerCamelCase__ ) if not return_dict: _UpperCAmelCase : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCamelCase__ , logits=lowerCamelCase__ , hidden_states=outputs.hidden_states )
322
0
'''simple docstring''' import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ = logging.get_logger(__name__) set_seed(770) lowerCamelCase__ = { 'c_attn': 'att_proj', 'c_proj': 'out_proj', 'c_fc': 'in_proj', 'transformer.': '', 'h.': 'layers.', 'ln_1': 'layernorm_1', 'ln_2': 'layernorm_2', 'ln_f': 'layernorm_final', 'wpe': 'position_embeds_layer', 'wte': 'input_embeds_layer', } lowerCamelCase__ = { 'text_small': { 'repo_id': 'suno/bark', 'file_name': 'text.pt', }, 'coarse_small': { 'repo_id': 'suno/bark', 'file_name': 'coarse.pt', }, 'fine_small': { 'repo_id': 'suno/bark', 'file_name': 'fine.pt', }, 'text': { 'repo_id': 'suno/bark', 'file_name': 'text_2.pt', }, 'coarse': { 'repo_id': 'suno/bark', 'file_name': 'coarse_2.pt', }, 'fine': { 'repo_id': 'suno/bark', 'file_name': 'fine_2.pt', }, } lowerCamelCase__ = os.path.dirname(os.path.abspath(__file__)) lowerCamelCase__ = os.path.join(os.path.expanduser('~'), '.cache') lowerCamelCase__ = os.path.join(os.getenv('XDG_CACHE_HOME', default_cache_dir), 'suno', 'bark_v0') def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=False ): _UpperCAmelCase : Any = model_type if use_small: key += "_small" return os.path.join(__lowerCAmelCase , REMOTE_MODEL_PATHS[key]["file_name"] ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) hf_hub_download(repo_id=__lowerCAmelCase , filename=__lowerCAmelCase , local_dir=__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase="text" ): if model_type == "text": _UpperCAmelCase : Union[str, Any] = BarkSemanticModel _UpperCAmelCase : Optional[int] = BarkSemanticConfig _UpperCAmelCase : List[Any] = BarkSemanticGenerationConfig elif model_type == "coarse": _UpperCAmelCase : Any = BarkCoarseModel _UpperCAmelCase : Dict = BarkCoarseConfig _UpperCAmelCase : Union[str, Any] = BarkCoarseGenerationConfig elif model_type == "fine": _UpperCAmelCase : List[Any] = BarkFineModel _UpperCAmelCase : List[Any] = BarkFineConfig _UpperCAmelCase : List[Any] = BarkFineGenerationConfig else: raise NotImplementedError() _UpperCAmelCase : Optional[Any] = F"""{model_type}_small""" if use_small else model_type _UpperCAmelCase : Optional[int] = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(__lowerCAmelCase ): logger.info(F"""{model_type} model not found, downloading into `{CACHE_DIR}`.""" ) _download(model_info["repo_id"] , model_info["file_name"] ) _UpperCAmelCase : int = torch.load(__lowerCAmelCase , map_location=__lowerCAmelCase ) # this is a hack _UpperCAmelCase : Optional[int] = checkpoint["model_args"] if "input_vocab_size" not in model_args: _UpperCAmelCase : Optional[int] = model_args["vocab_size"] _UpperCAmelCase : str = model_args["vocab_size"] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments _UpperCAmelCase : Optional[Any] = model_args.pop("n_head" ) _UpperCAmelCase : Tuple = model_args.pop("n_embd" ) _UpperCAmelCase : Tuple = model_args.pop("n_layer" ) _UpperCAmelCase : Optional[Any] = ConfigClass(**checkpoint["model_args"] ) _UpperCAmelCase : str = ModelClass(config=__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = GenerationConfigClass() _UpperCAmelCase : List[Any] = model_generation_config _UpperCAmelCase : Any = checkpoint["model"] # fixup checkpoint _UpperCAmelCase : List[str] = "_orig_mod." for k, v in list(state_dict.items() ): if k.startswith(__lowerCAmelCase ): # replace part of the key with corresponding layer name in HF implementation _UpperCAmelCase : Optional[Any] = k[len(__lowerCAmelCase ) :] for old_layer_name in new_layer_name_dict: _UpperCAmelCase : int = new_k.replace(__lowerCAmelCase , new_layer_name_dict[old_layer_name] ) _UpperCAmelCase : Optional[int] = state_dict.pop(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = set(state_dict.keys() ) - set(model.state_dict().keys() ) _UpperCAmelCase : List[Any] = {k for k in extra_keys if not k.endswith(".attn.bias" )} _UpperCAmelCase : Union[str, Any] = set(model.state_dict().keys() ) - set(state_dict.keys() ) _UpperCAmelCase : Tuple = {k for k in missing_keys if not k.endswith(".attn.bias" )} if len(__lowerCAmelCase ) != 0: raise ValueError(F"""extra keys found: {extra_keys}""" ) if len(__lowerCAmelCase ) != 0: raise ValueError(F"""missing keys: {missing_keys}""" ) model.load_state_dict(__lowerCAmelCase , strict=__lowerCAmelCase ) _UpperCAmelCase : Dict = model.num_parameters(exclude_embeddings=__lowerCAmelCase ) _UpperCAmelCase : str = checkpoint["best_val_loss"].item() logger.info(F"""model loaded: {round(n_params/1e6 , 1 )}M params, {round(__lowerCAmelCase , 3 )} loss""" ) model.eval() model.to(__lowerCAmelCase ) del checkpoint, state_dict return model def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase="text" ): if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() _UpperCAmelCase : Any = "cpu" # do conversion on cpu _UpperCAmelCase : Optional[Any] = _get_ckpt_path(__lowerCAmelCase , use_small=__lowerCAmelCase ) _UpperCAmelCase : Tuple = _load_model(__lowerCAmelCase , __lowerCAmelCase , model_type=__lowerCAmelCase , use_small=__lowerCAmelCase ) # load bark initial model _UpperCAmelCase : int = _bark_load_model(__lowerCAmelCase , "cpu" , model_type=__lowerCAmelCase , use_small=__lowerCAmelCase ) if model_type == "text": _UpperCAmelCase : Union[str, Any] = bark_model["model"] if model.num_parameters(exclude_embeddings=__lowerCAmelCase ) != bark_model.get_num_params(): raise ValueError("initial and new models don't have the same number of parameters" ) # check if same output as the bark model _UpperCAmelCase : Any = 5 _UpperCAmelCase : List[str] = 10 if model_type in ["text", "coarse"]: _UpperCAmelCase : List[str] = torch.randint(256 , (batch_size, sequence_length) , dtype=torch.int ) _UpperCAmelCase : Any = bark_model(__lowerCAmelCase )[0] _UpperCAmelCase : Optional[Any] = model(__lowerCAmelCase ) # take last logits _UpperCAmelCase : Optional[int] = output_new_model_total.logits[:, [-1], :] else: _UpperCAmelCase : Union[str, Any] = 3 _UpperCAmelCase : List[str] = 8 _UpperCAmelCase : Union[str, Any] = torch.randint(256 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) _UpperCAmelCase : str = model(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Any = bark_model(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : List[Any] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("initial and new outputs don't have the same shape" ) if (output_new_model - output_old_model).abs().max().item() > 1e-3: raise ValueError("initial and new outputs are not equal" ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) model.save_pretrained(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): _UpperCAmelCase : Union[str, Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = BarkSemanticConfig.from_pretrained(os.path.join(__lowerCAmelCase , "config.json" ) ) _UpperCAmelCase : Union[str, Any] = BarkCoarseConfig.from_pretrained(os.path.join(__lowerCAmelCase , "config.json" ) ) _UpperCAmelCase : Union[str, Any] = BarkFineConfig.from_pretrained(os.path.join(__lowerCAmelCase , "config.json" ) ) _UpperCAmelCase : List[str] = EncodecConfig.from_pretrained("facebook/encodec_24khz" ) _UpperCAmelCase : Any = BarkSemanticModel.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : Dict = BarkCoarseModel.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : Tuple = BarkFineModel.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : Any = EncodecModel.from_pretrained("facebook/encodec_24khz" ) _UpperCAmelCase : Tuple = BarkConfig.from_sub_model_configs( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Any = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) _UpperCAmelCase : int = BarkModel(__lowerCAmelCase ) _UpperCAmelCase : int = semantic _UpperCAmelCase : List[Any] = coarseAcoustic _UpperCAmelCase : str = fineAcoustic _UpperCAmelCase : Union[str, Any] = codec _UpperCAmelCase : List[Any] = bark_generation_config Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) bark.save_pretrained(__lowerCAmelCase , repo_id=__lowerCAmelCase , push_to_hub=__lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('model_type', type=str, help='text, coarse or fine.') parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--is_small', action='store_true', help='convert the small version instead of the large.') lowerCamelCase__ = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
359
'''simple docstring''' import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def __lowerCAmelCase (__lowerCAmelCase ): if is_torch_version("<" , "2.0.0" ) or not hasattr(__lowerCAmelCase , "_dynamo" ): return False return isinstance(__lowerCAmelCase , torch._dynamo.eval_frame.OptimizedModule ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase = True ): _UpperCAmelCase : Any = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) _UpperCAmelCase : Dict = is_compiled_module(__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : Optional[int] = model _UpperCAmelCase : Any = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = model.module if not keep_fpaa_wrapper: _UpperCAmelCase : List[Any] = getattr(__lowerCAmelCase , "forward" ) _UpperCAmelCase : Dict = model.__dict__.pop("_original_forward" , __lowerCAmelCase ) if original_forward is not None: while hasattr(__lowerCAmelCase , "__wrapped__" ): _UpperCAmelCase : Optional[int] = forward.__wrapped__ if forward == original_forward: break _UpperCAmelCase : Dict = forward if getattr(__lowerCAmelCase , "_converted_to_transformer_engine" , __lowerCAmelCase ): convert_model(__lowerCAmelCase , to_transformer_engine=__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : int = model _UpperCAmelCase : str = compiled_model return model def __lowerCAmelCase (): PartialState().wait_for_everyone() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(__lowerCAmelCase , __lowerCAmelCase ) elif PartialState().local_process_index == 0: torch.save(__lowerCAmelCase , __lowerCAmelCase ) @contextmanager def __lowerCAmelCase (**__lowerCAmelCase ): for key, value in kwargs.items(): _UpperCAmelCase : str = str(__lowerCAmelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def __lowerCAmelCase (__lowerCAmelCase ): if not hasattr(__lowerCAmelCase , "__qualname__" ) and not hasattr(__lowerCAmelCase , "__name__" ): _UpperCAmelCase : List[str] = getattr(__lowerCAmelCase , "__class__" , __lowerCAmelCase ) if hasattr(__lowerCAmelCase , "__qualname__" ): return obj.__qualname__ if hasattr(__lowerCAmelCase , "__name__" ): return obj.__name__ return str(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): for key, value in source.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = destination.setdefault(__lowerCAmelCase , {} ) merge_dicts(__lowerCAmelCase , __lowerCAmelCase ) else: _UpperCAmelCase : Optional[int] = value return destination def __lowerCAmelCase (__lowerCAmelCase = None ): if port is None: _UpperCAmelCase : Tuple = 29_500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
322
0
'''simple docstring''' import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format='%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s', datefmt='%Y-%m-%d %H:%M:%S', level=os.environ.get('LOGLEVEL', 'INFO').upper(), stream=sys.stdout, ) lowerCamelCase__ = logging.getLogger(__name__) lowerCamelCase__ = {'facebook/bart-base': BartForConditionalGeneration} lowerCamelCase__ = {'facebook/bart-base': BartTokenizer} def __lowerCAmelCase (): _UpperCAmelCase : List[Any] = argparse.ArgumentParser(description="Export Bart model + Beam Search to ONNX graph." ) parser.add_argument( "--validation_file" , type=__lowerCAmelCase , default=__lowerCAmelCase , help="A csv or a json file containing the validation data." ) parser.add_argument( "--max_length" , type=__lowerCAmelCase , default=5 , help="The maximum total input sequence length after tokenization." , ) parser.add_argument( "--num_beams" , type=__lowerCAmelCase , default=__lowerCAmelCase , help=( "Number of beams to use for evaluation. This argument will be " "passed to ``model.generate``, which is used during ``evaluate`` and ``predict``." ) , ) parser.add_argument( "--model_name_or_path" , type=__lowerCAmelCase , help="Path to pretrained model or model identifier from huggingface.co/models." , required=__lowerCAmelCase , ) parser.add_argument( "--config_name" , type=__lowerCAmelCase , default=__lowerCAmelCase , help="Pretrained config name or path if not the same as model_name" , ) parser.add_argument( "--device" , type=__lowerCAmelCase , default="cpu" , help="Device where the model will be run" , ) parser.add_argument("--output_file_path" , type=__lowerCAmelCase , default=__lowerCAmelCase , help="Where to store the final ONNX file." ) _UpperCAmelCase : Tuple = parser.parse_args() return args def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase="cpu" ): _UpperCAmelCase : List[str] = model_dict[model_name].from_pretrained(__lowerCAmelCase ).to(__lowerCAmelCase ) _UpperCAmelCase : Any = tokenizer_dict[model_name].from_pretrained(__lowerCAmelCase ) if model_name in ["facebook/bart-base"]: _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : List[str] = None _UpperCAmelCase : List[Any] = 0 return huggingface_model, tokenizer def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): model.eval() _UpperCAmelCase : Optional[Any] = None _UpperCAmelCase : Optional[Any] = torch.jit.script(BARTBeamSearchGenerator(__lowerCAmelCase ) ) with torch.no_grad(): _UpperCAmelCase : int = "My friends are cool but they eat too many carbs." _UpperCAmelCase : List[Any] = tokenizer([ARTICLE_TO_SUMMARIZE] , max_length=1_024 , return_tensors="pt" ).to(model.device ) _UpperCAmelCase : int = model.generate( inputs["input_ids"] , attention_mask=inputs["attention_mask"] , num_beams=__lowerCAmelCase , max_length=__lowerCAmelCase , early_stopping=__lowerCAmelCase , decoder_start_token_id=model.config.decoder_start_token_id , ) torch.onnx.export( __lowerCAmelCase , ( inputs["input_ids"], inputs["attention_mask"], num_beams, max_length, model.config.decoder_start_token_id, ) , __lowerCAmelCase , opset_version=14 , input_names=["input_ids", "attention_mask", "num_beams", "max_length", "decoder_start_token_id"] , output_names=["output_ids"] , dynamic_axes={ "input_ids": {0: "batch", 1: "seq"}, "output_ids": {0: "batch", 1: "seq_out"}, } , example_outputs=__lowerCAmelCase , ) logger.info("Model exported to {}".format(__lowerCAmelCase ) ) _UpperCAmelCase : Tuple = remove_dup_initializers(os.path.abspath(__lowerCAmelCase ) ) logger.info("Deduplicated and optimized model written to {}".format(__lowerCAmelCase ) ) _UpperCAmelCase : Optional[Any] = onnxruntime.InferenceSession(__lowerCAmelCase ) _UpperCAmelCase : Tuple = ort_sess.run( __lowerCAmelCase , { "input_ids": inputs["input_ids"].cpu().numpy(), "attention_mask": inputs["attention_mask"].cpu().numpy(), "num_beams": np.array(__lowerCAmelCase ), "max_length": np.array(__lowerCAmelCase ), "decoder_start_token_id": np.array(model.config.decoder_start_token_id ), } , ) np.testing.assert_allclose(summary_ids.cpu().numpy() , ort_out[0] , rtol=1e-3 , atol=1e-3 ) logger.info("Model outputs from torch and ONNX Runtime are similar." ) logger.info("Success." ) def __lowerCAmelCase (): _UpperCAmelCase : List[str] = parse_args() _UpperCAmelCase : List[Any] = 5 _UpperCAmelCase : Optional[int] = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.setLevel(logging.INFO ) transformers.utils.logging.set_verbosity_error() _UpperCAmelCase : List[Any] = torch.device(args.device ) _UpperCAmelCase : Union[str, Any] = load_model_tokenizer(args.model_name_or_path , __lowerCAmelCase ) if model.config.decoder_start_token_id is None: raise ValueError("Make sure that `config.decoder_start_token_id` is correctly defined" ) model.to(__lowerCAmelCase ) if args.max_length: _UpperCAmelCase : Tuple = args.max_length if args.num_beams: _UpperCAmelCase : Optional[Any] = args.num_beams if args.output_file_path: _UpperCAmelCase : Optional[Any] = args.output_file_path else: _UpperCAmelCase : Optional[Any] = "BART.onnx" logger.info("Exporting model to ONNX" ) export_and_validate_model(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) if __name__ == "__main__": main()
360
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): if number < 0: raise ValueError("number must not be negative" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
322
0
'''simple docstring''' from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowerCamelCase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __lowerCAmelCase (__lowerCAmelCase ): if isinstance(__lowerCAmelCase , torch.Tensor ): return image elif isinstance(__lowerCAmelCase , PIL.Image.Image ): _UpperCAmelCase : int = [image] _UpperCAmelCase : str = [trans(img.convert("RGB" ) ) for img in image] _UpperCAmelCase : Optional[Any] = torch.stack(__lowerCAmelCase ) return image class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Tuple , lowerCamelCase__ : int , lowerCamelCase__ : int ) ->int: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM _UpperCAmelCase : Tuple = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : str ) ->Union[str, Any]: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(F"""The value of strength should in [0.0, 1.0] but is {strength}""" ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Dict , lowerCamelCase__ : List[str] , lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = min(int(num_inference_steps * strength ) , lowerCamelCase__ ) _UpperCAmelCase : str = max(num_inference_steps - init_timestep , 0 ) _UpperCAmelCase : List[str] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : str , lowerCamelCase__ : str , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[Any]=None ) ->str: '''simple docstring''' if not isinstance(lowerCamelCase__ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowerCamelCase__ )}""" ) _UpperCAmelCase : Union[str, Any] = image.to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(lowerCamelCase__ )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) _UpperCAmelCase : List[str] = init_latents.shape _UpperCAmelCase : Optional[int] = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=lowerCamelCase__ , dtype=lowerCamelCase__ ) # get latents print("add noise to latents at timestep" , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.scheduler.add_noise(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = init_latents return latents @torch.no_grad() def __call__( self : Any , lowerCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image] = None , lowerCamelCase__ : float = 0.8 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase__ : float = 0.0 , lowerCamelCase__ : int = 50 , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , ) ->Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowerCamelCase__ ) # 2. Preprocess image _UpperCAmelCase : Dict = preprocess(lowerCamelCase__ ) # 3. set timesteps self.scheduler.set_timesteps(lowerCamelCase__ , device=self.device ) _UpperCAmelCase : Any = self.get_timesteps(lowerCamelCase__ , lowerCamelCase__ , self.device ) _UpperCAmelCase : List[Any] = timesteps[:1].repeat(lowerCamelCase__ ) # 4. Prepare latent variables _UpperCAmelCase : Optional[int] = self.prepare_latents(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , self.unet.dtype , self.device , lowerCamelCase__ ) _UpperCAmelCase : Any = latents # 5. Denoising loop for t in self.progress_bar(lowerCamelCase__ ): # 1. predict noise model_output _UpperCAmelCase : Union[str, Any] = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _UpperCAmelCase : int = self.scheduler.step( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , eta=lowerCamelCase__ , use_clipped_model_output=lowerCamelCase__ , generator=lowerCamelCase__ , ).prev_sample _UpperCAmelCase : Dict = (image / 2 + 0.5).clamp(0 , 1 ) _UpperCAmelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCAmelCase : str = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowerCamelCase__ )
361
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('Program to check whether a number is a Perfect number or not...') lowerCamelCase__ = int(input('Enter number: ').strip()) print(F'''{number} is {"" if perfect(number) else "not "}a Perfect Number.''')
322
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class lowerCAmelCase__ ( unittest.TestCase ): lowerCAmelCase : Union[str, Any] = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCAmelCase : str = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : str ) ->str: '''simple docstring''' _UpperCAmelCase : Any = TextaTextGenerationPipeline(model=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) return generator, ["Something to write", "Something else"] def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : int ) ->List[str]: '''simple docstring''' _UpperCAmelCase : str = generator("Something there" ) self.assertEqual(lowerCamelCase__ , [{"generated_text": ANY(lowerCamelCase__ )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]["generated_text"].startswith("Something there" ) ) _UpperCAmelCase : Dict = generator(["This is great !", "Something else"] , num_return_sequences=2 , do_sample=lowerCamelCase__ ) self.assertEqual( lowerCamelCase__ , [ [{"generated_text": ANY(lowerCamelCase__ )}, {"generated_text": ANY(lowerCamelCase__ )}], [{"generated_text": ANY(lowerCamelCase__ )}, {"generated_text": ANY(lowerCamelCase__ )}], ] , ) _UpperCAmelCase : Union[str, Any] = generator( ["This is great !", "Something else"] , num_return_sequences=2 , batch_size=2 , do_sample=lowerCamelCase__ ) self.assertEqual( lowerCamelCase__ , [ [{"generated_text": ANY(lowerCamelCase__ )}, {"generated_text": ANY(lowerCamelCase__ )}], [{"generated_text": ANY(lowerCamelCase__ )}, {"generated_text": ANY(lowerCamelCase__ )}], ] , ) with self.assertRaises(lowerCamelCase__ ): generator(4 ) @require_torch def lowerCAmelCase__ ( self : Any ) ->str: '''simple docstring''' _UpperCAmelCase : int = pipeline("text2text-generation" , model="patrickvonplaten/t5-tiny-random" , framework="pt" ) # do_sample=False necessary for reproducibility _UpperCAmelCase : Union[str, Any] = generator("Something there" , do_sample=lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , [{"generated_text": ""}] ) _UpperCAmelCase : List[str] = 3 _UpperCAmelCase : Dict = generator( "Something there" , num_return_sequences=lowerCamelCase__ , num_beams=lowerCamelCase__ , ) _UpperCAmelCase : List[str] = [ {"generated_text": "Beide Beide Beide Beide Beide Beide Beide Beide Beide"}, {"generated_text": "Beide Beide Beide Beide Beide Beide Beide Beide"}, {"generated_text": ""}, ] self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = generator("This is a test" , do_sample=lowerCamelCase__ , num_return_sequences=2 , return_tensors=lowerCamelCase__ ) self.assertEqual( lowerCamelCase__ , [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ] , ) _UpperCAmelCase : List[str] = generator.model.config.eos_token_id _UpperCAmelCase : str = "<pad>" _UpperCAmelCase : int = generator( ["This is a test", "This is a second test"] , do_sample=lowerCamelCase__ , num_return_sequences=2 , batch_size=2 , return_tensors=lowerCamelCase__ , ) self.assertEqual( lowerCamelCase__ , [ [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ], [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ], ] , ) @require_tf def lowerCAmelCase__ ( self : Any ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Dict = pipeline("text2text-generation" , model="patrickvonplaten/t5-tiny-random" , framework="tf" ) # do_sample=False necessary for reproducibility _UpperCAmelCase : Any = generator("Something there" , do_sample=lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , [{"generated_text": ""}] )
362
'''simple docstring''' from collections.abc import Sequence def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return sum(c * (x**i) for i, c in enumerate(__lowerCAmelCase ) ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Dict = 0.0 for coeff in reversed(__lowerCAmelCase ): _UpperCAmelCase : int = result * x + coeff return result if __name__ == "__main__": lowerCamelCase__ = (0.0, 0.0, 5.0, 9.3, 7.0) lowerCamelCase__ = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
322
0
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class lowerCAmelCase__ ( UpperCAmelCase__ ): # to overwrite at feature extractactor specific tests lowerCAmelCase : Optional[Any] = None lowerCAmelCase : List[str] = None @property def lowerCAmelCase__ ( self : Tuple ) ->str: '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def lowerCAmelCase__ ( self : Dict ) ->Tuple: '''simple docstring''' _UpperCAmelCase : int = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(lowerCamelCase__ , "feature_size" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "sampling_rate" ) ) self.assertTrue(hasattr(lowerCamelCase__ , "padding_value" ) ) def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _UpperCAmelCase : int = self.feature_extraction_class(**self.feat_extract_dict ) _UpperCAmelCase : Optional[Any] = feat_extract.model_input_names[0] _UpperCAmelCase : int = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) for x, y in zip(lowerCamelCase__ , processed_features[input_name] ) ) ) _UpperCAmelCase : List[str] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=lowerCamelCase__ ) _UpperCAmelCase : int = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) _UpperCAmelCase : Any = processed_features[input_name] if len(batch_features_input.shape ) < 3: _UpperCAmelCase : List[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def lowerCAmelCase__ ( self : int ) ->Dict: '''simple docstring''' _UpperCAmelCase : Tuple = self.feat_extract_tester.prepare_inputs_for_common(equal_length=lowerCamelCase__ ) _UpperCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _UpperCAmelCase : Union[str, Any] = feat_extract.model_input_names[0] _UpperCAmelCase : Union[str, Any] = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) _UpperCAmelCase : Dict = processed_features[input_name] if len(batch_features_input.shape ) < 3: _UpperCAmelCase : Tuple = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def lowerCAmelCase__ ( self : str ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = self.feat_extract_tester.prepare_inputs_for_common(equal_length=lowerCamelCase__ ) _UpperCAmelCase : Any = self.feature_extraction_class(**self.feat_extract_dict ) _UpperCAmelCase : Optional[Any] = feat_extract.model_input_names[0] _UpperCAmelCase : Dict = BatchFeature({input_name: speech_inputs} , tensor_type="tf" ) _UpperCAmelCase : Tuple = processed_features[input_name] if len(batch_features_input.shape ) < 3: _UpperCAmelCase : List[str] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : str=False ) ->Optional[Any]: '''simple docstring''' def _inputs_have_equal_length(lowerCamelCase__ : int ): _UpperCAmelCase : Optional[int] = len(input[0] ) for input_slice in input[1:]: if len(lowerCamelCase__ ) != length: return False return True def _inputs_are_equal(lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] ): if len(lowerCamelCase__ ) != len(lowerCamelCase__ ): return False for input_slice_a, input_slice_a in zip(lowerCamelCase__ , lowerCamelCase__ ): if not np.allclose(np.asarray(lowerCamelCase__ ) , np.asarray(lowerCamelCase__ ) , atol=1E-3 ): return False return True _UpperCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _UpperCAmelCase : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(numpify=lowerCamelCase__ ) _UpperCAmelCase : Any = feat_extract.model_input_names[0] _UpperCAmelCase : Union[str, Any] = BatchFeature({input_name: speech_inputs} ) _UpperCAmelCase : List[str] = self.feat_extract_tester.seq_length_diff _UpperCAmelCase : Optional[int] = self.feat_extract_tester.max_seq_length + pad_diff _UpperCAmelCase : Union[str, Any] = self.feat_extract_tester.min_seq_length _UpperCAmelCase : str = self.feat_extract_tester.batch_size _UpperCAmelCase : str = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _UpperCAmelCase : Optional[Any] = feat_extract.pad(lowerCamelCase__ , padding=lowerCamelCase__ ) _UpperCAmelCase : int = input_a[input_name] _UpperCAmelCase : Tuple = feat_extract.pad(lowerCamelCase__ , padding="longest" ) _UpperCAmelCase : Any = input_a[input_name] _UpperCAmelCase : Any = feat_extract.pad(lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[-1] ) ) _UpperCAmelCase : List[Any] = input_a[input_name] _UpperCAmelCase : List[str] = feat_extract.pad(lowerCamelCase__ , padding="longest" , return_tensors="np" ) _UpperCAmelCase : Any = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(lowerCamelCase__ ): feat_extract.pad(lowerCamelCase__ , padding="max_length" )[input_name] _UpperCAmelCase : Any = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=lowerCamelCase__ , return_tensors="np" ) _UpperCAmelCase : Dict = input_a[input_name] self.assertFalse(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertTrue(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertTrue(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertTrue(_inputs_are_equal(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _UpperCAmelCase : Dict = feat_extract.pad(lowerCamelCase__ , pad_to_multiple_of=10 ) _UpperCAmelCase : List[Any] = input_a[input_name] _UpperCAmelCase : Tuple = feat_extract.pad(lowerCamelCase__ , padding="longest" , pad_to_multiple_of=10 ) _UpperCAmelCase : Any = input_a[input_name] _UpperCAmelCase : Union[str, Any] = feat_extract.pad( lowerCamelCase__ , padding="max_length" , pad_to_multiple_of=10 , max_length=lowerCamelCase__ ) _UpperCAmelCase : int = input_a[input_name] _UpperCAmelCase : str = feat_extract.pad( lowerCamelCase__ , padding="max_length" , pad_to_multiple_of=10 , max_length=lowerCamelCase__ , return_tensors="np" , ) _UpperCAmelCase : Dict = input_a[input_name] self.assertTrue(all(len(lowerCamelCase__ ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(lowerCamelCase__ , lowerCamelCase__ ) ) _UpperCAmelCase : Dict = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(lowerCamelCase__ ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _UpperCAmelCase : Dict = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Any=False ) ->Optional[Any]: '''simple docstring''' def _inputs_have_equal_length(lowerCamelCase__ : Optional[int] ): _UpperCAmelCase : Optional[Any] = len(input[0] ) for input_slice in input[1:]: if len(lowerCamelCase__ ) != length: return False return True def _inputs_are_equal(lowerCamelCase__ : Tuple , lowerCamelCase__ : str ): if len(lowerCamelCase__ ) != len(lowerCamelCase__ ): return False for input_slice_a, input_slice_a in zip(lowerCamelCase__ , lowerCamelCase__ ): if not np.allclose(np.asarray(lowerCamelCase__ ) , np.asarray(lowerCamelCase__ ) , atol=1E-3 ): return False return True _UpperCAmelCase : Any = self.feature_extraction_class(**self.feat_extract_dict ) _UpperCAmelCase : Any = self.feat_extract_tester.prepare_inputs_for_common(numpify=lowerCamelCase__ ) _UpperCAmelCase : Tuple = feat_extract.model_input_names[0] _UpperCAmelCase : Any = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _UpperCAmelCase : Tuple = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[0] ) , truncation=lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = input_a[input_name] _UpperCAmelCase : str = feat_extract.pad(lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[0] ) ) _UpperCAmelCase : Optional[int] = input_a[input_name] self.assertTrue(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertFalse(_inputs_have_equal_length(lowerCamelCase__ ) ) # truncate to smallest with np _UpperCAmelCase : List[str] = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[0] ) , return_tensors="np" , truncation=lowerCamelCase__ , ) _UpperCAmelCase : Any = input_a[input_name] _UpperCAmelCase : List[Any] = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[0] ) , return_tensors="np" ) _UpperCAmelCase : List[Any] = input_a[input_name] self.assertTrue(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(lowerCamelCase__ ) ) # truncate to middle _UpperCAmelCase : Any = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[1] ) , truncation=lowerCamelCase__ , return_tensors="np" , ) _UpperCAmelCase : Any = input_a[input_name] _UpperCAmelCase : List[Any] = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[1] ) , truncation=lowerCamelCase__ ) _UpperCAmelCase : List[str] = input_a[input_name] _UpperCAmelCase : Union[str, Any] = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[1] ) , return_tensors="np" ) _UpperCAmelCase : Optional[int] = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertTrue(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertTrue(_inputs_are_equal(lowerCamelCase__ , lowerCamelCase__ ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(lowerCamelCase__ ): feat_extract.pad(lowerCamelCase__ , truncation=lowerCamelCase__ )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(lowerCamelCase__ ): feat_extract.pad(lowerCamelCase__ , padding="longest" , truncation=lowerCamelCase__ )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(lowerCamelCase__ ): feat_extract.pad(lowerCamelCase__ , padding="longest" , truncation=lowerCamelCase__ )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(lowerCamelCase__ ): feat_extract.pad(lowerCamelCase__ , padding="max_length" , truncation=lowerCamelCase__ )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _UpperCAmelCase : Any = 12 _UpperCAmelCase : Dict = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=lowerCamelCase__ , truncation=lowerCamelCase__ , ) _UpperCAmelCase : Optional[int] = input_a[input_name] _UpperCAmelCase : Optional[Any] = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=lowerCamelCase__ , ) _UpperCAmelCase : int = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _UpperCAmelCase : Dict = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _UpperCAmelCase : Any = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(lowerCamelCase__ ) ) self.assertFalse(_inputs_have_equal_length(lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Any: '''simple docstring''' self._check_padding(numpify=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' self._check_padding(numpify=lowerCamelCase__ ) def lowerCAmelCase__ ( self : str ) ->int: '''simple docstring''' self._check_truncation(numpify=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] ) ->List[Any]: '''simple docstring''' self._check_truncation(numpify=lowerCamelCase__ ) @require_torch def lowerCAmelCase__ ( self : Union[str, Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _UpperCAmelCase : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _UpperCAmelCase : Tuple = feat_extract.model_input_names[0] _UpperCAmelCase : List[Any] = BatchFeature({input_name: speech_inputs} ) _UpperCAmelCase : int = feat_extract.pad(lowerCamelCase__ , padding="longest" , return_tensors="np" )[input_name] _UpperCAmelCase : Dict = feat_extract.pad(lowerCamelCase__ , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def lowerCAmelCase__ ( self : int ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_dict ) _UpperCAmelCase : Optional[int] = self.feat_extract_tester.prepare_inputs_for_common() _UpperCAmelCase : Any = feat_extract.model_input_names[0] _UpperCAmelCase : Tuple = BatchFeature({input_name: speech_inputs} ) _UpperCAmelCase : Optional[int] = feat_extract.pad(lowerCamelCase__ , padding="longest" , return_tensors="np" )[input_name] _UpperCAmelCase : Union[str, Any] = feat_extract.pad(lowerCamelCase__ , padding="longest" , return_tensors="tf" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Any = self.feat_extract_dict _UpperCAmelCase : List[Any] = True _UpperCAmelCase : Any = self.feature_extraction_class(**lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _UpperCAmelCase : Dict = [len(lowerCamelCase__ ) for x in speech_inputs] _UpperCAmelCase : int = feat_extract.model_input_names[0] _UpperCAmelCase : Any = BatchFeature({input_name: speech_inputs} ) _UpperCAmelCase : Dict = feat_extract.pad(lowerCamelCase__ , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , lowerCamelCase__ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , lowerCamelCase__ ) def lowerCAmelCase__ ( self : int ) ->int: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.feat_extract_dict _UpperCAmelCase : Dict = True _UpperCAmelCase : List[Any] = self.feature_extraction_class(**lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _UpperCAmelCase : List[Any] = [len(lowerCamelCase__ ) for x in speech_inputs] _UpperCAmelCase : List[str] = feat_extract.model_input_names[0] _UpperCAmelCase : Union[str, Any] = BatchFeature({input_name: speech_inputs} ) _UpperCAmelCase : List[str] = min(lowerCamelCase__ ) _UpperCAmelCase : Tuple = feat_extract.pad( lowerCamelCase__ , padding="max_length" , max_length=lowerCamelCase__ , truncation=lowerCamelCase__ , return_tensors="np" ) self.assertIn("attention_mask" , lowerCamelCase__ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
363
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[Any] = len(__lowerCAmelCase ) _UpperCAmelCase : Tuple = sum(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _UpperCAmelCase : Any = True for i in range(1 , s + 1 ): _UpperCAmelCase : List[Any] = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): _UpperCAmelCase : Optional[int] = dp[i][j - 1] if arr[i - 1] <= j: _UpperCAmelCase : Any = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: _UpperCAmelCase : List[Any] = s - 2 * j break return diff
322
0
'''simple docstring''' from __future__ import annotations def __lowerCAmelCase (__lowerCAmelCase ): if len(__lowerCAmelCase ) < 2: raise ValueError("Monogons and Digons are not polygons in the Euclidean space" ) if any(i <= 0 for i in nums ): raise ValueError("All values must be greater than 0" ) _UpperCAmelCase : int = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
364
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'microsoft/resnet-50': 'https://huggingface.co/microsoft/resnet-50/blob/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): lowerCAmelCase : int = "resnet" lowerCAmelCase : Union[str, Any] = ["basic", "bottleneck"] def __init__( self : Dict , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : Any=64 , lowerCamelCase__ : Optional[int]=[2_56, 5_12, 10_24, 20_48] , lowerCamelCase__ : int=[3, 4, 6, 3] , lowerCamelCase__ : Dict="bottleneck" , lowerCamelCase__ : Dict="relu" , lowerCamelCase__ : List[Any]=False , lowerCamelCase__ : Any=None , lowerCamelCase__ : int=None , **lowerCamelCase__ : Tuple , ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) _UpperCAmelCase : str = num_channels _UpperCAmelCase : List[str] = embedding_size _UpperCAmelCase : Tuple = hidden_sizes _UpperCAmelCase : Dict = depths _UpperCAmelCase : List[Any] = layer_type _UpperCAmelCase : Optional[int] = hidden_act _UpperCAmelCase : Tuple = downsample_in_first_stage _UpperCAmelCase : str = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(lowerCamelCase__ ) + 1 )] _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = get_aligned_output_features_output_indices( out_features=lowerCamelCase__ , out_indices=lowerCamelCase__ , stage_names=self.stage_names ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = version.parse("1.11" ) @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowerCAmelCase__ ( self : str ) ->float: '''simple docstring''' return 1E-3
322
0
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-classification/requirements.txt') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "rb" ) as f: _UpperCAmelCase : List[str] = Image.open(__lowerCAmelCase ) return im.convert("RGB" ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={ "help": "Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub)." } , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the training data."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the validation data."} ) lowerCAmelCase : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default="google/vit-base-patch16-224-in21k" , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCAmelCase__ )} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) lowerCAmelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCAmelCase : str = field(default=UpperCAmelCase__ , metadata={"help": "Name or path of preprocessor config."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = torch.stack([example["pixel_values"] for example in examples] ) _UpperCAmelCase : Tuple = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def __lowerCAmelCase (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , __lowerCAmelCase , __lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(__lowerCAmelCase ) transformers.utils.logging.set_verbosity(__lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _UpperCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCAmelCase : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCAmelCase : List[Any] = {} if data_args.train_dir is not None: _UpperCAmelCase : str = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCAmelCase : Optional[Any] = os.path.join(data_args.validation_dir , "**" ) _UpperCAmelCase : Any = load_dataset( "imagefolder" , data_files=__lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCAmelCase : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __lowerCAmelCase ) and data_args.train_val_split > 0.0: _UpperCAmelCase : List[Any] = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCAmelCase : List[str] = split["train"] _UpperCAmelCase : Union[str, Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase : Optional[int] = dataset["train"].features["labels"].names _UpperCAmelCase : int = {}, {} for i, label in enumerate(__lowerCAmelCase ): _UpperCAmelCase : int = str(__lowerCAmelCase ) _UpperCAmelCase : str = label # Load the accuracy metric from the datasets package _UpperCAmelCase : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(__lowerCAmelCase ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowerCAmelCase ) , labelaid=__lowerCAmelCase , idalabel=__lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCAmelCase : Dict = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCAmelCase : int = image_processor.size["shortest_edge"] else: _UpperCAmelCase : int = (image_processor.size["height"], image_processor.size["width"]) _UpperCAmelCase : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCAmelCase : Optional[int] = Compose( [ RandomResizedCrop(__lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCAmelCase : Union[str, Any] = Compose( [ Resize(__lowerCAmelCase ), CenterCrop(__lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCAmelCase : Dict = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(__lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCAmelCase : Optional[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(__lowerCAmelCase ) # Initalize our trainer _UpperCAmelCase : Union[str, Any] = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=__lowerCAmelCase , tokenizer=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : Any = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : int = last_checkpoint _UpperCAmelCase : Dict = trainer.train(resume_from_checkpoint=__lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowerCAmelCase ) trainer.save_metrics("eval" , __lowerCAmelCase ) # Write model card and (optionally) push to hub _UpperCAmelCase : int = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowerCAmelCase ) else: trainer.create_model_card(**__lowerCAmelCase ) if __name__ == "__main__": main()
365
'''simple docstring''' from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowerCamelCase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __lowerCAmelCase (__lowerCAmelCase ): if isinstance(__lowerCAmelCase , torch.Tensor ): return image elif isinstance(__lowerCAmelCase , PIL.Image.Image ): _UpperCAmelCase : int = [image] _UpperCAmelCase : str = [trans(img.convert("RGB" ) ) for img in image] _UpperCAmelCase : Optional[Any] = torch.stack(__lowerCAmelCase ) return image class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Tuple , lowerCamelCase__ : int , lowerCamelCase__ : int ) ->int: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM _UpperCAmelCase : Tuple = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : str ) ->Union[str, Any]: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(F"""The value of strength should in [0.0, 1.0] but is {strength}""" ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Dict , lowerCamelCase__ : List[str] , lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = min(int(num_inference_steps * strength ) , lowerCamelCase__ ) _UpperCAmelCase : str = max(num_inference_steps - init_timestep , 0 ) _UpperCAmelCase : List[str] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : str , lowerCamelCase__ : str , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[Any]=None ) ->str: '''simple docstring''' if not isinstance(lowerCamelCase__ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowerCamelCase__ )}""" ) _UpperCAmelCase : Union[str, Any] = image.to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(lowerCamelCase__ )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) _UpperCAmelCase : List[str] = init_latents.shape _UpperCAmelCase : Optional[int] = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=lowerCamelCase__ , dtype=lowerCamelCase__ ) # get latents print("add noise to latents at timestep" , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.scheduler.add_noise(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = init_latents return latents @torch.no_grad() def __call__( self : Any , lowerCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image] = None , lowerCamelCase__ : float = 0.8 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase__ : float = 0.0 , lowerCamelCase__ : int = 50 , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , ) ->Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowerCamelCase__ ) # 2. Preprocess image _UpperCAmelCase : Dict = preprocess(lowerCamelCase__ ) # 3. set timesteps self.scheduler.set_timesteps(lowerCamelCase__ , device=self.device ) _UpperCAmelCase , _UpperCAmelCase : Any = self.get_timesteps(lowerCamelCase__ , lowerCamelCase__ , self.device ) _UpperCAmelCase : List[Any] = timesteps[:1].repeat(lowerCamelCase__ ) # 4. Prepare latent variables _UpperCAmelCase : Optional[int] = self.prepare_latents(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , self.unet.dtype , self.device , lowerCamelCase__ ) _UpperCAmelCase : Any = latents # 5. Denoising loop for t in self.progress_bar(lowerCamelCase__ ): # 1. predict noise model_output _UpperCAmelCase : Union[str, Any] = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _UpperCAmelCase : int = self.scheduler.step( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , eta=lowerCamelCase__ , use_clipped_model_output=lowerCamelCase__ , generator=lowerCamelCase__ , ).prev_sample _UpperCAmelCase : Dict = (image / 2 + 0.5).clamp(0 , 1 ) _UpperCAmelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCAmelCase : str = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowerCamelCase__ )
322
0
'''simple docstring''' import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput lowerCamelCase__ = 'scheduler_config.json' class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[int] = 1 lowerCAmelCase : Any = 2 lowerCAmelCase : Union[str, Any] = 3 lowerCAmelCase : Union[str, Any] = 4 lowerCAmelCase : Union[str, Any] = 5 @dataclass class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : jnp.ndarray class lowerCAmelCase__ : lowerCAmelCase : Any = SCHEDULER_CONFIG_NAME lowerCAmelCase : int = ["dtype"] lowerCAmelCase : Dict = [] lowerCAmelCase : Optional[Any] = True @classmethod def lowerCAmelCase__ ( cls : List[Any] , lowerCamelCase__ : Dict[str, Any] = None , lowerCamelCase__ : Optional[str] = None , lowerCamelCase__ : List[str]=False , **lowerCamelCase__ : str , ) ->Any: '''simple docstring''' _UpperCAmelCase : int = cls.load_config( pretrained_model_name_or_path=lowerCamelCase__ , subfolder=lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ , **lowerCamelCase__ , ) _UpperCAmelCase : Union[str, Any] = cls.from_config(lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ , **lowerCamelCase__ ) if hasattr(lowerCamelCase__ , "create_state" ) and getattr(lowerCamelCase__ , "has_state" , lowerCamelCase__ ): _UpperCAmelCase : List[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Union[str, os.PathLike] , lowerCamelCase__ : bool = False , **lowerCamelCase__ : Dict ) ->Optional[int]: '''simple docstring''' self.save_config(save_directory=lowerCamelCase__ , push_to_hub=lowerCamelCase__ , **lowerCamelCase__ ) @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Tuple: '''simple docstring''' return self._get_compatibles() @classmethod def lowerCAmelCase__ ( cls : Dict ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = list(set([cls.__name__] + cls._compatibles ) ) _UpperCAmelCase : Optional[int] = importlib.import_module(__name__.split("." )[0] ) _UpperCAmelCase : Union[str, Any] = [ getattr(lowerCamelCase__ , lowerCamelCase__ ) for c in compatible_classes_str if hasattr(lowerCamelCase__ , lowerCamelCase__ ) ] return compatible_classes def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): assert len(__lowerCAmelCase ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(__lowerCAmelCase ) - x.ndim) ) , __lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=0.9_9_9 , __lowerCAmelCase=jnp.floataa ): def alpha_bar(__lowerCAmelCase ): return math.cos((time_step + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 _UpperCAmelCase : Optional[Any] = [] for i in range(__lowerCAmelCase ): _UpperCAmelCase : List[Any] = i / num_diffusion_timesteps _UpperCAmelCase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(__lowerCAmelCase ) / alpha_bar(__lowerCAmelCase ) , __lowerCAmelCase ) ) return jnp.array(__lowerCAmelCase , dtype=__lowerCAmelCase ) @flax.struct.dataclass class lowerCAmelCase__ : lowerCAmelCase : jnp.ndarray lowerCAmelCase : jnp.ndarray lowerCAmelCase : jnp.ndarray @classmethod def lowerCAmelCase__ ( cls : Tuple , lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : int = scheduler.config if config.trained_betas is not None: _UpperCAmelCase : Dict = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": _UpperCAmelCase : Any = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _UpperCAmelCase : int = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _UpperCAmelCase : Any = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( F"""beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}""" ) _UpperCAmelCase : Tuple = 1.0 - betas _UpperCAmelCase : Tuple = jnp.cumprod(lowerCamelCase__ , axis=0 ) return cls( alphas=lowerCamelCase__ , betas=lowerCamelCase__ , alphas_cumprod=lowerCamelCase__ , ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[int] = state.alphas_cumprod _UpperCAmelCase : List[Any] = alphas_cumprod[timesteps] ** 0.5 _UpperCAmelCase : Any = sqrt_alpha_prod.flatten() _UpperCAmelCase : Union[str, Any] = broadcast_to_shape_from_left(__lowerCAmelCase , original_samples.shape ) _UpperCAmelCase : str = (1 - alphas_cumprod[timesteps]) ** 0.5 _UpperCAmelCase : List[str] = sqrt_one_minus_alpha_prod.flatten() _UpperCAmelCase : str = broadcast_to_shape_from_left(__lowerCAmelCase , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[int] = get_sqrt_alpha_prod(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : int = get_sqrt_alpha_prod(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
366
'''simple docstring''' from __future__ import annotations from collections.abc import Callable lowerCamelCase__ = list[list[float | int]] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(size + 1 )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : float for row in range(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = matrix[row][col] _UpperCAmelCase : Optional[int] = vector[row][0] _UpperCAmelCase : int = 0 _UpperCAmelCase : Union[str, Any] = 0 while row < size and col < size: # pivoting _UpperCAmelCase : Optional[Any] = max((abs(augmented[rowa][col] ), rowa) for rowa in range(__lowerCAmelCase , __lowerCAmelCase ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: _UpperCAmelCase , _UpperCAmelCase : str = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , __lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[rowa][col] / augmented[row][col] _UpperCAmelCase : Optional[Any] = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , __lowerCAmelCase ): for row in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[row][col] / augmented[col][col] for cola in range(__lowerCAmelCase , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(__lowerCAmelCase ) ] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(__lowerCAmelCase )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix = [[0] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int for x_val, y_val in enumerate(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = (x_val + 1) ** (size - col - 1) _UpperCAmelCase : int = y_val _UpperCAmelCase : List[str] = solve(__lowerCAmelCase , __lowerCAmelCase ) def interpolated_func(__lowerCAmelCase ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(__lowerCAmelCase ) ) return interpolated_func def __lowerCAmelCase (__lowerCAmelCase ): return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def __lowerCAmelCase (__lowerCAmelCase = question_function , __lowerCAmelCase = 10 ): _UpperCAmelCase : list[int] = [func(__lowerCAmelCase ) for x_val in range(1 , order + 1 )] _UpperCAmelCase : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] _UpperCAmelCase : int = 0 _UpperCAmelCase : Callable[[int], int] _UpperCAmelCase : int for poly in polynomials: _UpperCAmelCase : int = 1 while func(__lowerCAmelCase ) == poly(__lowerCAmelCase ): x_val += 1 ret += poly(__lowerCAmelCase ) return ret if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : Optional[int] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = ["a", "b", "c"] # Defaults to last layer if both are None _UpperCAmelCase : str = get_aligned_output_features_output_indices(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , ["c"] ) self.assertEqual(lowerCamelCase__ , [2] ) # Out indices set to match out features _UpperCAmelCase : Union[str, Any] = get_aligned_output_features_output_indices(["a", "c"] , lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , ["a", "c"] ) self.assertEqual(lowerCamelCase__ , [0, 2] ) # Out features set to match out indices _UpperCAmelCase : Dict = get_aligned_output_features_output_indices(lowerCamelCase__ , [0, 2] , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , ["a", "c"] ) self.assertEqual(lowerCamelCase__ , [0, 2] ) # Out features selected from negative indices _UpperCAmelCase : str = get_aligned_output_features_output_indices(lowerCamelCase__ , [-3, -1] , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , ["a", "c"] ) self.assertEqual(lowerCamelCase__ , [-3, -1] ) def lowerCAmelCase__ ( self : List[str] ) ->str: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(["a", "b"] , (0, 1) , lowerCamelCase__ ) # Out features must be a list with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(("a", "b") , (0, 1) , ["a", "b"] ) # Out features must be a subset of stage names with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(["a", "b"] , (0, 1) , ["a"] ) # Out indices must be a list or tuple with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(lowerCamelCase__ , 0 , ["a", "b"] ) # Out indices must be a subset of stage names with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(lowerCamelCase__ , (0, 1) , ["a"] ) # Out features and out indices must be the same length with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(["a", "b"] , (0,) , ["a", "b", "c"] ) # Out features should match out indices with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(["a", "b"] , (0, 2) , ["a", "b", "c"] ) # Out features and out indices should be in order with self.assertRaises(lowerCamelCase__ ): verify_out_features_out_indices(["b", "a"] , (0, 1) , ["a", "b"] ) # Check passes with valid inputs verify_out_features_out_indices(["a", "b", "d"] , (0, 1, -1) , ["a", "b", "c", "d"] ) def lowerCAmelCase__ ( self : Optional[Any] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Any = BackboneMixin() _UpperCAmelCase : str = ["a", "b", "c"] _UpperCAmelCase : Optional[int] = ["a", "c"] _UpperCAmelCase : Optional[int] = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly _UpperCAmelCase : Optional[int] = ["a", "b"] self.assertEqual(backbone.out_features , ["a", "b"] ) self.assertEqual(backbone.out_indices , [0, 1] ) _UpperCAmelCase : List[Any] = [-3, -1] self.assertEqual(backbone.out_features , ["a", "c"] ) self.assertEqual(backbone.out_indices , [-3, -1] )
367
'''simple docstring''' from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
322
0
'''simple docstring''' import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument( '--original_config_file', type=str, required=True, help='The YAML config file corresponding to the original architecture.', ) parser.add_argument( '--num_in_channels', default=None, type=int, help='The number of input channels. If `None` number of input channels will be automatically inferred.', ) parser.add_argument( '--image_size', default=512, type=int, help=( 'The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2' ' Base. Use 768 for Stable Diffusion v2.' ), ) parser.add_argument( '--extract_ema', action='store_true', help=( 'Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights' ' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield' ' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.' ), ) parser.add_argument( '--upcast_attention', action='store_true', help=( 'Whether the attention computation should always be upcasted. This is necessary when running stable' ' diffusion 2.1.' ), ) parser.add_argument( '--from_safetensors', action='store_true', help='If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.', ) parser.add_argument( '--to_safetensors', action='store_true', help='Whether to store pipeline in safetensors format or not.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--device', type=str, help='Device to use (e.g. cpu, cuda:0, cuda:1, etc.)') def __lowerCAmelCase (__lowerCAmelCase ): if string == "True": return True elif string == "False": return False else: raise ValueError(F"""could not parse string as bool {string}""" ) parser.add_argument( '--use_linear_projection', help='Override for use linear projection', required=False, type=parse_bool ) parser.add_argument('--cross_attention_dim', help='Override for cross attention_dim', required=False, type=int) lowerCamelCase__ = parser.parse_args() lowerCamelCase__ = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
368
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar lowerCamelCase__ = TypeVar('T') class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Union[str, Any] , lowerCamelCase__ : T ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Dict = data _UpperCAmelCase : Node[T] | None = None def __str__( self : Any ) ->str: '''simple docstring''' return F"""{self.data}""" class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Tuple ) ->None: '''simple docstring''' _UpperCAmelCase : Node[T] | None = None def __iter__( self : List[str] ) ->Iterator[T]: '''simple docstring''' _UpperCAmelCase : Any = self.top while node: yield node.data _UpperCAmelCase : Dict = node.next def __str__( self : Dict ) ->str: '''simple docstring''' return "->".join([str(lowerCamelCase__ ) for item in self] ) def __len__( self : Optional[int] ) ->int: '''simple docstring''' return len(tuple(iter(self ) ) ) def lowerCAmelCase__ ( self : List[Any] ) ->bool: '''simple docstring''' return self.top is None def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : T ) ->None: '''simple docstring''' _UpperCAmelCase : List[Any] = Node(lowerCamelCase__ ) if not self.is_empty(): _UpperCAmelCase : Tuple = self.top _UpperCAmelCase : List[str] = node def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("pop from empty stack" ) assert isinstance(self.top , lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.top _UpperCAmelCase : Optional[Any] = self.top.next return pop_node.data def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("peek from empty stack" ) assert self.top is not None return self.top.data def lowerCAmelCase__ ( self : List[Any] ) ->None: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = None if __name__ == "__main__": from doctest import testmod testmod()
322
0
'''simple docstring''' import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): # Load configuration defined in the metadata file with open(__lowerCAmelCase ) as metadata_file: _UpperCAmelCase : List[str] = json.load(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = LukeConfig(use_entity_aware_attention=__lowerCAmelCase , **metadata["model_config"] ) # Load in the weights from the checkpoint_path _UpperCAmelCase : Tuple = torch.load(__lowerCAmelCase , map_location="cpu" ) # Load the entity vocab file _UpperCAmelCase : List[Any] = load_entity_vocab(__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = RobertaTokenizer.from_pretrained(metadata["model_config"]["bert_model_name"] ) # Add special tokens to the token vocabulary for downstream tasks _UpperCAmelCase : Dict = AddedToken("<ent>" , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = AddedToken("<ent2>" , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) tokenizer.add_special_tokens({"additional_special_tokens": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(__lowerCAmelCase ) with open(os.path.join(__lowerCAmelCase , LukeTokenizer.vocab_files_names["entity_vocab_file"] ) , "w" ) as f: json.dump(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Any = LukeTokenizer.from_pretrained(__lowerCAmelCase ) # Initialize the embeddings of the special tokens _UpperCAmelCase : List[Any] = state_dict["embeddings.word_embeddings.weight"] _UpperCAmelCase : Any = word_emb[tokenizer.convert_tokens_to_ids(["@"] )[0]].unsqueeze(0 ) _UpperCAmelCase : Union[str, Any] = word_emb[tokenizer.convert_tokens_to_ids(["#"] )[0]].unsqueeze(0 ) _UpperCAmelCase : Tuple = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _UpperCAmelCase : int = F"""encoder.layer.{layer_index}.attention.self.""" _UpperCAmelCase : Optional[Any] = state_dict[prefix + matrix_name] _UpperCAmelCase : str = state_dict[prefix + matrix_name] _UpperCAmelCase : List[Any] = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _UpperCAmelCase : List[Any] = state_dict["entity_embeddings.entity_embeddings.weight"] _UpperCAmelCase : Union[str, Any] = entity_emb[entity_vocab["[MASK]"]] _UpperCAmelCase : Optional[Any] = LukeModel(config=__lowerCAmelCase ).eval() _UpperCAmelCase : Union[str, Any] = model.load_state_dict(__lowerCAmelCase , strict=__lowerCAmelCase ) if not (len(__lowerCAmelCase ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {', '.join(__lowerCAmelCase )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith("entity_predictions" ) or key.startswith("lm_head" ) for key in unexpected_keys )): raise ValueError( "Unexpected keys" F""" {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}""" ) # Check outputs _UpperCAmelCase : Optional[Any] = LukeTokenizer.from_pretrained(__lowerCAmelCase , task="entity_classification" ) _UpperCAmelCase : List[Any] = ( "Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the" " new world number one avoid a humiliating second- round exit at Wimbledon ." ) _UpperCAmelCase : Tuple = (39, 42) _UpperCAmelCase : str = tokenizer(__lowerCAmelCase , entity_spans=[span] , add_prefix_space=__lowerCAmelCase , return_tensors="pt" ) _UpperCAmelCase : List[str] = model(**__lowerCAmelCase ) # Verify word hidden states if model_size == "large": _UpperCAmelCase : Any = torch.Size((1, 42, 1_024) ) _UpperCAmelCase : Union[str, Any] = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _UpperCAmelCase : Tuple = torch.Size((1, 42, 768) ) _UpperCAmelCase : Optional[Any] = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowerCAmelCase , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _UpperCAmelCase : Tuple = torch.Size((1, 1, 1_024) ) _UpperCAmelCase : Optional[int] = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _UpperCAmelCase : Any = torch.Size((1, 1, 768) ) _UpperCAmelCase : Optional[Any] = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __lowerCAmelCase , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print("Saving PyTorch model to {}".format(__lowerCAmelCase ) ) model.save_pretrained(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Any = {} with open(__lowerCAmelCase , "r" , encoding="utf-8" ) as f: for index, line in enumerate(__lowerCAmelCase ): _UpperCAmelCase : Tuple = line.rstrip().split("\t" ) _UpperCAmelCase : Dict = index return entity_vocab if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--checkpoint_path', type=str, help='Path to a pytorch_model.bin file.') parser.add_argument( '--metadata_path', default=None, type=str, help='Path to a metadata.json file, defining the configuration.' ) parser.add_argument( '--entity_vocab_path', default=None, type=str, help='Path to an entity_vocab.tsv file, containing the entity vocabulary.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to where to dump the output PyTorch model.' ) parser.add_argument( '--model_size', default='base', type=str, choices=['base', 'large'], help='Size of the model to be converted.' ) lowerCamelCase__ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
369
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : int = "speech_to_text_2" lowerCAmelCase : str = ["past_key_values"] lowerCAmelCase : int = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[Any] , lowerCamelCase__ : Tuple=1_00_00 , lowerCamelCase__ : Any=6 , lowerCamelCase__ : Tuple=20_48 , lowerCamelCase__ : List[Any]=4 , lowerCamelCase__ : Tuple=0.0 , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Tuple="relu" , lowerCamelCase__ : Dict=2_56 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : List[Any]=0.0 , lowerCamelCase__ : Optional[int]=0.0 , lowerCamelCase__ : List[Any]=0.0_2 , lowerCamelCase__ : Tuple=2 , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Any=1 , lowerCamelCase__ : int=0 , lowerCamelCase__ : str=2 , lowerCamelCase__ : List[Any]=10_24 , **lowerCamelCase__ : str , ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = vocab_size _UpperCAmelCase : Optional[int] = d_model _UpperCAmelCase : List[Any] = decoder_ffn_dim _UpperCAmelCase : Any = decoder_layers _UpperCAmelCase : int = decoder_attention_heads _UpperCAmelCase : Any = dropout _UpperCAmelCase : List[Any] = attention_dropout _UpperCAmelCase : Optional[int] = activation_dropout _UpperCAmelCase : List[Any] = activation_function _UpperCAmelCase : int = init_std _UpperCAmelCase : Dict = decoder_layerdrop _UpperCAmelCase : str = use_cache _UpperCAmelCase : Union[str, Any] = decoder_layers _UpperCAmelCase : Optional[Any] = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCAmelCase : Any = max_target_positions super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
322
0
'''simple docstring''' import re def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = re.compile( R"^(?:0|94|\+94|0{2}94)" R"7(0|1|2|4|5|6|7|8)" R"(-| |)" R"\d{7}$" ) return bool(re.search(__lowerCAmelCase , __lowerCAmelCase ) ) if __name__ == "__main__": lowerCamelCase__ = '0094702343221' print(is_sri_lankan_phone_number(phone))
370
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = 'cuda' if torch.cuda.is_available() else 'cpu' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=100 , __lowerCAmelCase=" " ): _UpperCAmelCase : Any = text.split(__lowerCAmelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(__lowerCAmelCase ) , __lowerCAmelCase )] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase , _UpperCAmelCase : Dict = [], [] for title, text in zip(documents["title"] , documents["text"] ): if text is not None: for passage in split_text(__lowerCAmelCase ): titles.append(title if title is not None else "" ) texts.append(__lowerCAmelCase ) return {"title": titles, "text": texts} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : str = ctx_tokenizer( documents["title"] , documents["text"] , truncation=__lowerCAmelCase , padding="longest" , return_tensors="pt" )["input_ids"] _UpperCAmelCase : str = ctx_encoder(input_ids.to(device=__lowerCAmelCase ) , return_dict=__lowerCAmelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): ###################################### logger.info("Step 1 - Create the dataset" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way _UpperCAmelCase : Optional[int] = load_dataset( "csv" , data_files=[rag_example_args.csv_path] , split="train" , delimiter="\t" , column_names=["title", "text"] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words _UpperCAmelCase : Optional[int] = dataset.map(__lowerCAmelCase , batched=__lowerCAmelCase , num_proc=processing_args.num_proc ) # And compute the embeddings _UpperCAmelCase : Union[str, Any] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) _UpperCAmelCase : Dict = Features( {"text": Value("string" ), "title": Value("string" ), "embeddings": Sequence(Value("float32" ) )} ) # optional, save as float32 instead of float64 to save space _UpperCAmelCase : int = dataset.map( partial(__lowerCAmelCase , ctx_encoder=__lowerCAmelCase , ctx_tokenizer=__lowerCAmelCase ) , batched=__lowerCAmelCase , batch_size=processing_args.batch_size , features=__lowerCAmelCase , ) # And finally save your dataset _UpperCAmelCase : List[Any] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset" ) dataset.save_to_disk(__lowerCAmelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("Step 2 - Index the dataset" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search _UpperCAmelCase : Any = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("embeddings" , custom_index=__lowerCAmelCase ) # And save the index _UpperCAmelCase : List[str] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset_hnsw_index.faiss" ) dataset.get_index("embeddings" ).save(__lowerCAmelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" / "my_knowledge_dataset.csv" ) , metadata={"help": "Path to a tab-separated csv file with columns 'title' and 'text'"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Question that is passed as input to RAG. Default is 'What does Moses' rod turn into ?'."} , ) lowerCAmelCase : str = field( default="facebook/rag-sequence-nq" , metadata={"help": "The RAG model to use. Either 'facebook/rag-sequence-nq' or 'facebook/rag-token-nq'"} , ) lowerCAmelCase : str = field( default="facebook/dpr-ctx_encoder-multiset-base" , metadata={ "help": ( "The DPR context encoder model to use. Either 'facebook/dpr-ctx_encoder-single-nq-base' or" " 'facebook/dpr-ctx_encoder-multiset-base'" ) } , ) lowerCAmelCase : Optional[str] = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" ) , metadata={"help": "Path to a directory where the dataset passages and the index will be saved"} , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": "The number of processes to use to split the documents into passages. Default is single process." } , ) lowerCAmelCase : int = field( default=16 , metadata={ "help": "The batch size to use when computing the passages embeddings using the DPR context encoder." } , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : int = field( default=768 , metadata={"help": "The dimension of the embeddings to pass to the HNSW Faiss index."} , ) lowerCAmelCase : int = field( default=128 , metadata={ "help": ( "The number of bi-directional links created for every new element during the HNSW index construction." ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
322
0
import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin lowerCamelCase__ = 1e-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class lowerCAmelCase__ : def __init__( self : Tuple , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any]=16 , lowerCamelCase__ : Any=13 , lowerCamelCase__ : Any=7 , lowerCamelCase__ : Any=14 , lowerCamelCase__ : Any=10 , lowerCamelCase__ : List[Any]=19 , lowerCamelCase__ : Optional[int]=5 , lowerCamelCase__ : Any=4 , lowerCamelCase__ : Dict=True , lowerCamelCase__ : str=16 , lowerCamelCase__ : Optional[Any]=2 , lowerCamelCase__ : str=4 , lowerCamelCase__ : str=4 , lowerCamelCase__ : Optional[Any]="gelu" , lowerCamelCase__ : Optional[Any]=0.1 , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Tuple=[1, 2, 3, 4, 5] , lowerCamelCase__ : Union[str, Any]=25 , lowerCamelCase__ : Tuple=5 , ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = d_model _UpperCAmelCase : Any = parent _UpperCAmelCase : Tuple = batch_size _UpperCAmelCase : Union[str, Any] = prediction_length _UpperCAmelCase : Dict = context_length _UpperCAmelCase : Tuple = cardinality _UpperCAmelCase : Dict = num_time_features _UpperCAmelCase : Any = lags_sequence _UpperCAmelCase : Optional[Any] = embedding_dimension _UpperCAmelCase : str = is_training _UpperCAmelCase : List[Any] = hidden_size _UpperCAmelCase : Optional[Any] = num_hidden_layers _UpperCAmelCase : List[str] = num_attention_heads _UpperCAmelCase : Union[str, Any] = intermediate_size _UpperCAmelCase : Dict = hidden_act _UpperCAmelCase : List[str] = hidden_dropout_prob _UpperCAmelCase : List[str] = attention_probs_dropout_prob _UpperCAmelCase : Optional[Any] = context_length _UpperCAmelCase : str = prediction_length + label_length _UpperCAmelCase : List[str] = label_length _UpperCAmelCase : Optional[Any] = moving_average _UpperCAmelCase : int = autocorrelation_factor def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[Any]: '''simple docstring''' return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : Any = config.context_length + max(config.lags_sequence ) _UpperCAmelCase : Dict = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) _UpperCAmelCase : Optional[Any] = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) _UpperCAmelCase : str = floats_tensor([self.batch_size, _past_length] ) _UpperCAmelCase : Optional[Any] = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs _UpperCAmelCase : List[str] = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) _UpperCAmelCase : int = floats_tensor([self.batch_size, config.prediction_length] ) _UpperCAmelCase : List[str] = { "past_values": past_values, "static_categorical_features": static_categorical_features, "past_time_features": past_time_features, "past_observed_mask": past_observed_mask, "future_time_features": future_time_features, "future_values": future_values, } return inputs_dict def lowerCAmelCase__ ( self : Any ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.get_config() _UpperCAmelCase : Any = self.prepare_autoformer_inputs_dict(lowerCamelCase__ ) return config, inputs_dict def lowerCAmelCase__ ( self : str ) ->List[str]: '''simple docstring''' _UpperCAmelCase : str = self.prepare_config_and_inputs() return config, inputs_dict def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Dict ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = AutoformerModel(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _UpperCAmelCase : List[Any] = model(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = outputs.encoder_last_hidden_state _UpperCAmelCase : str = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: _UpperCAmelCase : int = model.get_encoder() encoder.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = AutoformerEncoder.from_pretrained(lowerCamelCase__ ).to(lowerCamelCase__ ) _UpperCAmelCase : int = model.create_network_inputs(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) _UpperCAmelCase : List[str] = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) _UpperCAmelCase : Tuple = encoder(inputs_embeds=lowerCamelCase__ )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3 ) _UpperCAmelCase : Union[str, Any] = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) _UpperCAmelCase : Dict = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) _UpperCAmelCase : Optional[int] = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) _UpperCAmelCase : Union[str, Any] = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: _UpperCAmelCase : List[str] = model.get_decoder() decoder.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : int = AutoformerDecoder.from_pretrained(lowerCamelCase__ ).to(lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = decoder( trend=lowerCamelCase__ , inputs_embeds=lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3 ) @require_torch class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : int = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () lowerCAmelCase : Union[str, Any] = (AutoformerForPrediction,) if is_torch_available() else () lowerCAmelCase : List[Any] = {"feature-extraction": AutoformerModel} if is_torch_available() else {} lowerCAmelCase : List[Any] = False lowerCAmelCase : int = False lowerCAmelCase : List[Any] = False lowerCAmelCase : Union[str, Any] = False lowerCAmelCase : List[Any] = False lowerCAmelCase : List[Any] = False def lowerCAmelCase__ ( self : Tuple ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = AutoformerModelTester(self ) _UpperCAmelCase : int = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def lowerCAmelCase__ ( self : str ) ->Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCAmelCase__ ( self : List[Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: _UpperCAmelCase : int = model_class(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = model_class.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertEqual(info["missing_keys"] , [] ) def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*lowerCamelCase__ ) @unittest.skip(reason="Model has no tokens embeddings" ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[Any]: '''simple docstring''' pass def lowerCAmelCase__ ( self : Optional[int] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : int = inspect.signature(getattr(lowerCamelCase__ , "forward" ) ) # The main input is the name of the argument after `self` _UpperCAmelCase : Union[str, Any] = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Any ) ->Tuple: '''simple docstring''' _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : Optional[Any] = model_class(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase : List[str] = [*signature.parameters.keys()] _UpperCAmelCase : Optional[int] = [ "past_values", "past_time_features", "past_observed_mask", "static_categorical_features", "static_real_features", "future_values", "future_time_features", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("future_observed_mask" ) expected_arg_names.extend( [ "decoder_attention_mask", "head_mask", "decoder_head_mask", "cross_attn_head_mask", "encoder_outputs", "past_key_values", "output_hidden_states", "output_attentions", "use_cache", "return_dict", ] ) self.assertListEqual(arg_names[: len(lowerCamelCase__ )] , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Tuple = True _UpperCAmelCase : Union[str, Any] = getattr(self.model_tester , "seq_length" , lowerCamelCase__ ) _UpperCAmelCase : Dict = getattr(self.model_tester , "decoder_seq_length" , lowerCamelCase__ ) _UpperCAmelCase : Dict = getattr(self.model_tester , "encoder_seq_length" , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = getattr(self.model_tester , "d_model" , lowerCamelCase__ ) _UpperCAmelCase : Dict = getattr(self.model_tester , "num_attention_heads" , lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = d_model // num_attention_heads for model_class in self.all_model_classes: _UpperCAmelCase : List[Any] = True _UpperCAmelCase : Tuple = False _UpperCAmelCase : Optional[int] = True _UpperCAmelCase : Union[str, Any] = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _UpperCAmelCase : int = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _UpperCAmelCase : int = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _UpperCAmelCase : Optional[Any] = True _UpperCAmelCase : List[str] = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _UpperCAmelCase : List[str] = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = outputs.encoder_attentions self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) _UpperCAmelCase : Optional[Any] = len(lowerCamelCase__ ) _UpperCAmelCase : Any = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) # decoder attentions _UpperCAmelCase : Tuple = outputs.decoder_attentions self.assertIsInstance(lowerCamelCase__ , (list, tuple) ) self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions _UpperCAmelCase : Union[str, Any] = outputs.cross_attentions self.assertIsInstance(lowerCamelCase__ , (list, tuple) ) self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine _UpperCAmelCase : Any = True _UpperCAmelCase : Tuple = True _UpperCAmelCase : Tuple = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _UpperCAmelCase : Optional[int] = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(out_len + 2 , len(lowerCamelCase__ ) ) _UpperCAmelCase : List[Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def lowerCAmelCase__ ( self : List[Any] ) ->int: '''simple docstring''' super().test_retain_grad_hidden_states_attentions() def __lowerCAmelCase (__lowerCAmelCase="train-batch.pt" ): _UpperCAmelCase : Optional[int] = hf_hub_download(repo_id="hf-internal-testing/tourism-monthly-batch" , filename=__lowerCAmelCase , repo_type="dataset" ) _UpperCAmelCase : str = torch.load(__lowerCAmelCase , map_location=__lowerCAmelCase ) return batch @require_torch @slow class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : List[Any] = AutoformerModel.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(lowerCamelCase__ ) _UpperCAmelCase : Any = prepare_batch() with torch.no_grad(): _UpperCAmelCase : Any = model( past_values=batch["past_values"] , past_time_features=batch["past_time_features"] , past_observed_mask=batch["past_observed_mask"] , static_categorical_features=batch["static_categorical_features"] , future_values=batch["future_values"] , future_time_features=batch["future_time_features"] , )[0] _UpperCAmelCase : Dict = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : Tuple = torch.tensor( [[0.3_5_9_3, -1.3_3_9_8, 0.6_3_3_0], [0.2_2_7_9, 1.5_3_9_6, -0.1_7_9_2], [0.0_4_5_0, 1.3_2_2_5, -0.2_3_3_5]] , device=lowerCamelCase__ ) self.assertTrue(torch.allclose(output[0, :3, :3] , lowerCamelCase__ , atol=lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = AutoformerForPrediction.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(lowerCamelCase__ ) _UpperCAmelCase : Dict = prepare_batch("val-batch.pt" ) with torch.no_grad(): _UpperCAmelCase : Optional[int] = model( past_values=batch["past_values"] , past_time_features=batch["past_time_features"] , past_observed_mask=batch["past_observed_mask"] , static_categorical_features=batch["static_categorical_features"] , ).encoder_last_hidden_state _UpperCAmelCase : Dict = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = torch.tensor( [[-0.0_7_3_4, -0.9_0_3_6, 0.8_3_5_8], [4.7_1_8_6, 2.4_1_1_3, 1.9_5_8_1], [1.7_9_5_3, 2.3_5_5_8, 1.2_9_7_0]] , device=lowerCamelCase__ ) self.assertTrue(torch.allclose(output[0, :3, :3] , lowerCamelCase__ , atol=lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Dict ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = AutoformerForPrediction.from_pretrained("huggingface/autoformer-tourism-monthly" ).to(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = prepare_batch("val-batch.pt" ) with torch.no_grad(): _UpperCAmelCase : Optional[Any] = model.generate( static_categorical_features=batch["static_categorical_features"] , past_time_features=batch["past_time_features"] , past_values=batch["past_values"] , future_time_features=batch["future_time_features"] , past_observed_mask=batch["past_observed_mask"] , ) _UpperCAmelCase : Tuple = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , lowerCamelCase__ ) _UpperCAmelCase : Tuple = torch.tensor([31_30.67_63, 40_56.52_93, 70_53.07_86] , device=lowerCamelCase__ ) _UpperCAmelCase : List[str] = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , lowerCamelCase__ , rtol=1E-1 ) )
371
'''simple docstring''' import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCamelCase__ = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class lowerCAmelCase__ ( unittest.TestCase ): @classmethod def lowerCAmelCase__ ( cls : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : Tuple = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def lowerCAmelCase__ ( cls : Union[str, Any] ) ->int: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def lowerCAmelCase__ ( self : int ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ , repo_id="test-config" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : Dict = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ , repo_id="valid_org/test-config-org" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : int = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' CustomConfig.register_for_auto_class() _UpperCAmelCase : int = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) _UpperCAmelCase : str = AutoConfig.from_pretrained(F"""{USER}/test-dynamic-config""" , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : List[str] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated _UpperCAmelCase : Any = c.n_embd + 1 # int _UpperCAmelCase : List[Any] = c.resid_pdrop + 1.0 # float _UpperCAmelCase : Tuple = not c.scale_attn_weights # bool _UpperCAmelCase : List[Any] = c.summary_type + "foo" # str c.update_from_string( F"""n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}""" ) self.assertEqual(lowerCamelCase__ , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(lowerCamelCase__ , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(lowerCamelCase__ , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(lowerCamelCase__ , c.summary_type , "mismatch for key: summary_type" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = PretrainedConfig() _UpperCAmelCase : Tuple = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase__ , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) _UpperCAmelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase__ , lowerCamelCase__ )] if len(lowerCamelCase__ ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" F""" {', '.join(lowerCamelCase__ )}.""" ) def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): # config is in subfolder, the following should not work without specifying the subfolder _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = mock.Mock() _UpperCAmelCase : List[str] = 5_00 _UpperCAmelCase : Dict = {} _UpperCAmelCase : Tuple = HTTPError _UpperCAmelCase : Any = {} # Download this model to make sure it's in the cache. _UpperCAmelCase : int = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=lowerCamelCase__ ) as mock_head: _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = AutoConfig.from_pretrained("bert-base-cased" ) _UpperCAmelCase : str = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 _UpperCAmelCase : Optional[int] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 _UpperCAmelCase : Dict = ["config.42.0.0.json"] _UpperCAmelCase : Union[str, Any] = 7_68 configuration.save_pretrained(lowerCamelCase__ ) shutil.move(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , os.path.join(lowerCamelCase__ , "config.42.0.0.json" ) ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = "hf-internal-testing/test-two-configs" import transformers as new_transformers _UpperCAmelCase : Any = "v4.0.0" _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers _UpperCAmelCase : List[Any] = "v3.0.0" _UpperCAmelCase : int = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(old_configuration.hidden_size , 7_68 )
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Any = 1 _UpperCAmelCase : Dict = 2 while i * i <= n: _UpperCAmelCase : Optional[Any] = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def __lowerCAmelCase (): _UpperCAmelCase : int = 1 _UpperCAmelCase : int = 1 while True: i += 1 t_num += i if count_divisors(__lowerCAmelCase ) > 500: break return t_num if __name__ == "__main__": print(solution())
350
'''simple docstring''' from manim import * class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : List[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Optional[Any] = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Dict = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("CPU" , font_size=24 ) _UpperCAmelCase : Any = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(1 )] _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("GPU" , font_size=24 ) _UpperCAmelCase : str = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) gpu.align_to(lowerCamelCase__ , lowerCamelCase__ ) gpu.set_x(gpu.get_x() - 1 ) self.add(lowerCamelCase__ ) _UpperCAmelCase : List[str] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Any = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Model" , font_size=24 ) _UpperCAmelCase : Tuple = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.play( Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , ) _UpperCAmelCase : int = MarkupText( F"""First, an empty model skeleton is loaded\ninto <span fgcolor='{YELLOW}'>memory</span> without using much RAM.""" , font_size=24 , ) _UpperCAmelCase : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ , run_time=2.5 ) , Write(lowerCamelCase__ ) , Write(lowerCamelCase__ ) ) self.add(lowerCamelCase__ ) _UpperCAmelCase : int = [] _UpperCAmelCase : List[str] = [] _UpperCAmelCase : Dict = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : int = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ , opacity=0.7 ) cpu_target.move_to(lowerCamelCase__ ) cpu_target.generate_target() _UpperCAmelCase : Dict = 0.4_6 / 4 _UpperCAmelCase : Any = 0.4_6 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowerCamelCase__ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=lowerCamelCase__ , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=lowerCamelCase__ , buff=0.0 ) cpu_targs.append(lowerCamelCase__ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(lowerCamelCase__ ) ) second_animations.append(MoveToTarget(lowerCamelCase__ , run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(*lowerCamelCase__ ) self.wait()
322
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'facebook/timesformer': 'https://huggingface.co/facebook/timesformer/resolve/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Dict = "timesformer" def __init__( self : int , lowerCamelCase__ : List[Any]=2_24 , lowerCamelCase__ : List[str]=16 , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : Dict=8 , lowerCamelCase__ : Optional[int]=7_68 , lowerCamelCase__ : Optional[Any]=12 , lowerCamelCase__ : List[Any]=12 , lowerCamelCase__ : Tuple=30_72 , lowerCamelCase__ : List[Any]="gelu" , lowerCamelCase__ : str=0.0 , lowerCamelCase__ : Optional[Any]=0.0 , lowerCamelCase__ : Tuple=0.0_2 , lowerCamelCase__ : int=1E-6 , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : List[Any]="divided_space_time" , lowerCamelCase__ : Optional[int]=0 , **lowerCamelCase__ : int , ) ->Any: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Dict = image_size _UpperCAmelCase : Any = patch_size _UpperCAmelCase : Dict = num_channels _UpperCAmelCase : Union[str, Any] = num_frames _UpperCAmelCase : Optional[int] = hidden_size _UpperCAmelCase : int = num_hidden_layers _UpperCAmelCase : Tuple = num_attention_heads _UpperCAmelCase : Optional[Any] = intermediate_size _UpperCAmelCase : Tuple = hidden_act _UpperCAmelCase : Optional[int] = hidden_dropout_prob _UpperCAmelCase : Optional[Any] = attention_probs_dropout_prob _UpperCAmelCase : List[Any] = initializer_range _UpperCAmelCase : Any = layer_norm_eps _UpperCAmelCase : List[str] = qkv_bias _UpperCAmelCase : str = attention_type _UpperCAmelCase : List[Any] = drop_path_rate
351
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1_024 , __lowerCAmelCase=1_024 , __lowerCAmelCase=False , **__lowerCAmelCase ): _UpperCAmelCase : Any = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : List[str] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="train" , **__lowerCAmelCase ) _UpperCAmelCase : Dict = tok.pad_token_id def get_lens(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = tqdm( DataLoader(__lowerCAmelCase , batch_size=512 , num_workers=8 , shuffle=__lowerCAmelCase , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) _UpperCAmelCase : List[str] = [] for batch in dl: _UpperCAmelCase : Any = batch["input_ids"].ne(__lowerCAmelCase ).sum(1 ).tolist() _UpperCAmelCase : Tuple = batch["labels"].ne(__lowerCAmelCase ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__lowerCAmelCase , __lowerCAmelCase ): max_lens.append(max(__lowerCAmelCase , __lowerCAmelCase ) ) else: max_lens.extend(__lowerCAmelCase ) return max_lens _UpperCAmelCase : Dict = get_lens(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="val" , **__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = get_lens(__lowerCAmelCase ) pickle_save(__lowerCAmelCase , train_ds.len_file ) pickle_save(__lowerCAmelCase , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
322
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available lowerCamelCase__ = {'tokenization_herbert': ['HerbertTokenizer']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ['HerbertTokenizerFast'] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
352
'''simple docstring''' import pytest lowerCamelCase__ = '__dummy_dataset1__' lowerCamelCase__ = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = dataset_loading_script_name _UpperCAmelCase : Any = tmp_path / "datasets" / script_name script_dir.mkdir(parents=__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = script_dir / F"""{script_name}.py""" with open(__lowerCAmelCase , "w" ) as f: f.write(__lowerCAmelCase ) return str(__lowerCAmelCase )
322
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) lowerCamelCase__ = { 'configuration_speecht5': [ 'SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP', 'SpeechT5Config', 'SpeechT5HifiGanConfig', ], 'feature_extraction_speecht5': ['SpeechT5FeatureExtractor'], 'processing_speecht5': ['SpeechT5Processor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ['SpeechT5Tokenizer'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ 'SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'SpeechT5ForSpeechToText', 'SpeechT5ForSpeechToSpeech', 'SpeechT5ForTextToSpeech', 'SpeechT5Model', 'SpeechT5PreTrainedModel', 'SpeechT5HifiGan', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
353
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase__ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCamelCase__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCamelCase__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCamelCase__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"] , reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ] , ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] ) ->int: '''simple docstring''' import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : int=0.9 , lowerCamelCase__ : Dict=3 , lowerCamelCase__ : Dict=0.5 ) ->Any: '''simple docstring''' if NLTK_VERSION >= version.Version("3.6.5" ): _UpperCAmelCase : Dict = [ meteor_score.single_meteor_score( word_tokenize(lowerCamelCase__ ) , word_tokenize(lowerCamelCase__ ) , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] else: _UpperCAmelCase : Optional[int] = [ meteor_score.single_meteor_score(lowerCamelCase__ , lowerCamelCase__ , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] return {"meteor": np.mean(lowerCamelCase__ )}
322
0
'''simple docstring''' import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowerCAmelCase__ : def __init__( self : Dict , lowerCamelCase__ : Any , lowerCamelCase__ : Optional[Any]=3 , lowerCamelCase__ : Dict=32 , lowerCamelCase__ : Any=3 , lowerCamelCase__ : List[Any]=10 , lowerCamelCase__ : List[Any]=[8, 16, 32, 64] , lowerCamelCase__ : Union[str, Any]=[1, 1, 2, 1] , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : Optional[Any]="relu" , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : List[str]=None , lowerCamelCase__ : Optional[int]=["stage2", "stage3", "stage4"] , lowerCamelCase__ : Optional[Any]=[2, 3, 4] , lowerCamelCase__ : Optional[int]=1 , ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = parent _UpperCAmelCase : Tuple = batch_size _UpperCAmelCase : Any = image_size _UpperCAmelCase : str = num_channels _UpperCAmelCase : List[Any] = embeddings_size _UpperCAmelCase : str = hidden_sizes _UpperCAmelCase : Dict = depths _UpperCAmelCase : Union[str, Any] = is_training _UpperCAmelCase : Tuple = use_labels _UpperCAmelCase : Tuple = hidden_act _UpperCAmelCase : Tuple = num_labels _UpperCAmelCase : Dict = scope _UpperCAmelCase : str = len(lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = out_features _UpperCAmelCase : Dict = out_indices _UpperCAmelCase : str = num_groups def lowerCAmelCase__ ( self : Any ) ->Tuple: '''simple docstring''' _UpperCAmelCase : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase : Optional[int] = None if self.use_labels: _UpperCAmelCase : Any = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase : Union[str, Any] = self.get_config() return config, pixel_values, labels def lowerCAmelCase__ ( self : Dict ) ->Any: '''simple docstring''' return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Optional[int] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Optional[int] = BitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Optional[int] = model(lowerCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Dict ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = self.num_labels _UpperCAmelCase : Optional[int] = BitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : List[str] = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : str , lowerCamelCase__ : str , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : int = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Tuple = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _UpperCAmelCase : int = None _UpperCAmelCase : List[Any] = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCAmelCase : Union[str, Any] = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def lowerCAmelCase__ ( self : Dict ) ->int: '''simple docstring''' _UpperCAmelCase : Optional[int] = self.prepare_config_and_inputs() _UpperCAmelCase : List[str] = config_and_inputs _UpperCAmelCase : Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : int = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () lowerCAmelCase : Dict = ( {"feature-extraction": BitModel, "image-classification": BitForImageClassification} if is_torch_available() else {} ) lowerCAmelCase : Optional[int] = False lowerCAmelCase : str = False lowerCAmelCase : Any = False lowerCAmelCase : Dict = False lowerCAmelCase : Dict = False def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Tuple = BitModelTester(self ) _UpperCAmelCase : int = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def lowerCAmelCase__ ( self : int ) ->Optional[Any]: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase__ ( self : List[str] ) ->str: '''simple docstring''' return @unittest.skip(reason="Bit does not output attentions" ) def lowerCAmelCase__ ( self : Dict ) ->Any: '''simple docstring''' pass @unittest.skip(reason="Bit does not use inputs_embeds" ) def lowerCAmelCase__ ( self : Tuple ) ->Tuple: '''simple docstring''' pass @unittest.skip(reason="Bit does not support input and output embeddings" ) def lowerCAmelCase__ ( self : Tuple ) ->str: '''simple docstring''' pass def lowerCAmelCase__ ( self : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : int = model_class(lowerCamelCase__ ) _UpperCAmelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase : int = [*signature.parameters.keys()] _UpperCAmelCase : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def lowerCAmelCase__ ( self : int ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def lowerCAmelCase__ ( self : int ) ->Tuple: '''simple docstring''' _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : Tuple = model_class(config=lowerCamelCase__ ) for name, module in model.named_modules(): if isinstance(lowerCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def lowerCAmelCase__ ( self : str ) ->List[Any]: '''simple docstring''' def check_hidden_states_output(lowerCamelCase__ : Any , lowerCamelCase__ : int , lowerCamelCase__ : Dict ): _UpperCAmelCase : Union[str, Any] = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _UpperCAmelCase : Any = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _UpperCAmelCase : Optional[int] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCAmelCase : List[Any] = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : List[str] = ["preactivation", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: _UpperCAmelCase : Optional[int] = layer_type _UpperCAmelCase : Any = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase : List[str] = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip(reason="Bit does not use feedforward chunking" ) def lowerCAmelCase__ ( self : str ) ->List[Any]: '''simple docstring''' pass def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' _UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : Optional[Any] ) ->Tuple: '''simple docstring''' for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase : List[Any] = BitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def __lowerCAmelCase (): _UpperCAmelCase : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCAmelCase__ ( unittest.TestCase ): @cached_property def lowerCAmelCase__ ( self : Union[str, Any] ) ->Tuple: '''simple docstring''' return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowerCAmelCase__ ( self : List[str] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase__ ) _UpperCAmelCase : Tuple = self.default_image_processor _UpperCAmelCase : List[str] = prepare_img() _UpperCAmelCase : Any = image_processor(images=lowerCamelCase__ , return_tensors="pt" ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _UpperCAmelCase : List[str] = model(**lowerCamelCase__ ) # verify the logits _UpperCAmelCase : Dict = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = torch.tensor([[-0.6_5_2_6, -0.5_2_6_3, -1.4_3_9_8]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1E-4 ) ) @require_torch class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : int = (BitBackbone,) if is_torch_available() else () lowerCAmelCase : Any = BitConfig lowerCAmelCase : int = False def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' _UpperCAmelCase : Dict = BitModelTester(self )
354
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline lowerCamelCase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : int ) ->str: '''simple docstring''' if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Union[str, Any] = [label.strip() for label in labels.split("," ) if label.strip()] return labels def __call__( self : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] ) ->str: '''simple docstring''' if len(lowerCamelCase__ ) == 0 or len(lowerCamelCase__ ) == 0: raise ValueError("You must include at least one label and at least one sequence." ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( "The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. " "Make sure the passed template includes formatting syntax such as {{}} where the label should go." ).format(lowerCamelCase__ ) ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Optional[Any] = [sequences] _UpperCAmelCase : int = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(lowerCamelCase__ )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(UpperCAmelCase__ ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Union[str, Any] , lowerCamelCase__ : Optional[Any]=ZeroShotClassificationArgumentHandler() , *lowerCamelCase__ : List[str] , **lowerCamelCase__ : Any ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = args_parser super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) if self.entailment_id == -1: logger.warning( "Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to " "-1. Define a descriptive label2id mapping in the model config to ensure correct outputs." ) @property def lowerCAmelCase__ ( self : Any ) ->Union[str, Any]: '''simple docstring''' for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("entail" ): return ind return -1 def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Tuple , lowerCamelCase__ : int=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : str=TruncationStrategy.ONLY_FIRST , **lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : int = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( "Tokenizer was not supporting padding necessary for zero-shot, attempting to use " " `pad_token=eos_token`" ) _UpperCAmelCase : Optional[Any] = self.tokenizer.eos_token try: _UpperCAmelCase : List[str] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , ) except Exception as e: if "too short" in str(lowerCamelCase__ ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. _UpperCAmelCase : List[Any] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def lowerCAmelCase__ ( self : int , **lowerCamelCase__ : Union[str, Any] ) ->Tuple: '''simple docstring''' if kwargs.get("multi_class" , lowerCamelCase__ ) is not None: _UpperCAmelCase : int = kwargs["multi_class"] logger.warning( "The `multi_class` argument has been deprecated and renamed to `multi_label`. " "`multi_class` will be removed in a future version of Transformers." ) _UpperCAmelCase : Dict = {} if "candidate_labels" in kwargs: _UpperCAmelCase : List[Any] = self._args_parser._parse_labels(kwargs["candidate_labels"] ) if "hypothesis_template" in kwargs: _UpperCAmelCase : Dict = kwargs["hypothesis_template"] _UpperCAmelCase : List[str] = {} if "multi_label" in kwargs: _UpperCAmelCase : Optional[Any] = kwargs["multi_label"] return preprocess_params, {}, postprocess_params def __call__( self : int , lowerCamelCase__ : Union[str, List[str]] , *lowerCamelCase__ : str , **lowerCamelCase__ : Optional[Any] , ) ->Optional[int]: '''simple docstring''' if len(lowerCamelCase__ ) == 0: pass elif len(lowerCamelCase__ ) == 1 and "candidate_labels" not in kwargs: _UpperCAmelCase : int = args[0] else: raise ValueError(F"""Unable to understand extra arguments {args}""" ) return super().__call__(lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Any=None , lowerCamelCase__ : str="This example is {}." ) ->Tuple: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Optional[int] = self._args_parser(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for i, (candidate_label, sequence_pair) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): _UpperCAmelCase : Optional[int] = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(lowerCamelCase__ ) - 1, **model_input, } def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Optional[int] ) ->int: '''simple docstring''' _UpperCAmelCase : Dict = inputs["candidate_label"] _UpperCAmelCase : Optional[int] = inputs["sequence"] _UpperCAmelCase : Dict = {k: inputs[k] for k in self.tokenizer.model_input_names} _UpperCAmelCase : List[Any] = self.model(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = { "candidate_label": candidate_label, "sequence": sequence, "is_last": inputs["is_last"], **outputs, } return model_outputs def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple=False ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = [outputs["candidate_label"] for outputs in model_outputs] _UpperCAmelCase : Any = [outputs["sequence"] for outputs in model_outputs] _UpperCAmelCase : Optional[int] = np.concatenate([output["logits"].numpy() for output in model_outputs] ) _UpperCAmelCase : Optional[Any] = logits.shape[0] _UpperCAmelCase : Any = len(lowerCamelCase__ ) _UpperCAmelCase : str = N // n _UpperCAmelCase : str = logits.reshape((num_sequences, n, -1) ) if multi_label or len(lowerCamelCase__ ) == 1: # softmax over the entailment vs. contradiction dim for each label independently _UpperCAmelCase : int = self.entailment_id _UpperCAmelCase : List[Any] = -1 if entailment_id == 0 else 0 _UpperCAmelCase : str = reshaped_outputs[..., [contradiction_id, entailment_id]] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : str = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels _UpperCAmelCase : int = reshaped_outputs[..., self.entailment_id] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
322
0
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[str]=13 , lowerCamelCase__ : Optional[Any]=7 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : int=99 , lowerCamelCase__ : int=32 , lowerCamelCase__ : List[str]=5 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : Optional[int]=37 , lowerCamelCase__ : Tuple="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Union[str, Any]=0.1 , lowerCamelCase__ : Optional[int]=5_12 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : str=2 , lowerCamelCase__ : Union[str, Any]=0.0_2 , lowerCamelCase__ : Tuple=4 , ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = parent _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Optional[int] = seq_length _UpperCAmelCase : int = is_training _UpperCAmelCase : Dict = use_attention_mask _UpperCAmelCase : Optional[Any] = use_token_type_ids _UpperCAmelCase : int = use_labels _UpperCAmelCase : Optional[int] = vocab_size _UpperCAmelCase : Any = hidden_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : List[Any] = num_attention_heads _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : int = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob _UpperCAmelCase : Union[str, Any] = max_position_embeddings _UpperCAmelCase : Tuple = type_vocab_size _UpperCAmelCase : List[Any] = type_sequence_label_size _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Dict = num_choices def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Dict = None if self.use_attention_mask: _UpperCAmelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCAmelCase : Union[str, Any] = None if self.use_token_type_ids: _UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCAmelCase : int = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() _UpperCAmelCase : List[Any] = config_and_inputs _UpperCAmelCase : str = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Optional[int] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase__ ( self : Optional[int] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = FlaxAlbertModelTester(self ) @slow def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCAmelCase : List[str] = model_class_name.from_pretrained("albert-base-v2" ) _UpperCAmelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = FlaxAlbertModel.from_pretrained("albert-base-v2" ) _UpperCAmelCase : List[Any] = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) _UpperCAmelCase : Optional[int] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCAmelCase : Dict = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )[0] _UpperCAmelCase : List[Any] = (1, 11, 7_68) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : str = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase__ , atol=1E-4 ) )
355
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase = 4_000_000 ): _UpperCAmelCase : List[Any] = [] _UpperCAmelCase , _UpperCAmelCase : Dict = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__lowerCAmelCase ) _UpperCAmelCase , _UpperCAmelCase : Any = b, a + b return sum(__lowerCAmelCase ) if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): def update_area_of_max_square(__lowerCAmelCase , __lowerCAmelCase ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 _UpperCAmelCase : Optional[Any] = update_area_of_max_square(__lowerCAmelCase , col + 1 ) _UpperCAmelCase : Union[str, Any] = update_area_of_max_square(row + 1 , col + 1 ) _UpperCAmelCase : str = update_area_of_max_square(row + 1 , __lowerCAmelCase ) if mat[row][col]: _UpperCAmelCase : str = 1 + min([right, diagonal, down] ) _UpperCAmelCase : Dict = max(largest_square_area[0] , __lowerCAmelCase ) return sub_problem_sol else: return 0 _UpperCAmelCase : Union[str, Any] = [0] update_area_of_max_square(0 , 0 ) return largest_square_area[0] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): def update_area_of_max_square_using_dp_array( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] _UpperCAmelCase : str = update_area_of_max_square_using_dp_array(__lowerCAmelCase , col + 1 , __lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , __lowerCAmelCase ) _UpperCAmelCase : Any = update_area_of_max_square_using_dp_array(row + 1 , __lowerCAmelCase , __lowerCAmelCase ) if mat[row][col]: _UpperCAmelCase : Any = 1 + min([right, diagonal, down] ) _UpperCAmelCase : List[str] = max(largest_square_area[0] , __lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = sub_problem_sol return sub_problem_sol else: return 0 _UpperCAmelCase : Optional[Any] = [0] _UpperCAmelCase : Union[str, Any] = [[-1] * cols for _ in range(__lowerCAmelCase )] update_area_of_max_square_using_dp_array(0 , 0 , __lowerCAmelCase ) return largest_square_area[0] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [[0] * (cols + 1) for _ in range(rows + 1 )] _UpperCAmelCase : Dict = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): _UpperCAmelCase : Union[str, Any] = dp_array[row][col + 1] _UpperCAmelCase : str = dp_array[row + 1][col + 1] _UpperCAmelCase : List[Any] = dp_array[row + 1][col] if mat[row][col] == 1: _UpperCAmelCase : Tuple = 1 + min(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : List[str] = max(dp_array[row][col] , __lowerCAmelCase ) else: _UpperCAmelCase : Optional[Any] = 0 return largest_square_area def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Tuple = [0] * (cols + 1) _UpperCAmelCase : Optional[Any] = [0] * (cols + 1) _UpperCAmelCase : str = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): _UpperCAmelCase : List[Any] = current_row[col + 1] _UpperCAmelCase : Union[str, Any] = next_row[col + 1] _UpperCAmelCase : List[str] = next_row[col] if mat[row][col] == 1: _UpperCAmelCase : Any = 1 + min(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : int = max(current_row[col] , __lowerCAmelCase ) else: _UpperCAmelCase : str = 0 _UpperCAmelCase : Optional[Any] = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
356
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[str]=13 , lowerCamelCase__ : Optional[Any]=7 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : int=99 , lowerCamelCase__ : int=32 , lowerCamelCase__ : List[str]=5 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : Optional[int]=37 , lowerCamelCase__ : Tuple="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Union[str, Any]=0.1 , lowerCamelCase__ : Optional[int]=5_12 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : str=2 , lowerCamelCase__ : Union[str, Any]=0.0_2 , lowerCamelCase__ : Tuple=4 , ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = parent _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Optional[int] = seq_length _UpperCAmelCase : int = is_training _UpperCAmelCase : Dict = use_attention_mask _UpperCAmelCase : Optional[Any] = use_token_type_ids _UpperCAmelCase : int = use_labels _UpperCAmelCase : Optional[int] = vocab_size _UpperCAmelCase : Any = hidden_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : List[Any] = num_attention_heads _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : int = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob _UpperCAmelCase : Union[str, Any] = max_position_embeddings _UpperCAmelCase : Tuple = type_vocab_size _UpperCAmelCase : List[Any] = type_sequence_label_size _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Dict = num_choices def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Dict = None if self.use_attention_mask: _UpperCAmelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCAmelCase : Union[str, Any] = None if self.use_token_type_ids: _UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCAmelCase : int = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = config_and_inputs _UpperCAmelCase : str = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Optional[int] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase__ ( self : Optional[int] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = FlaxAlbertModelTester(self ) @slow def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCAmelCase : List[str] = model_class_name.from_pretrained("albert-base-v2" ) _UpperCAmelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = FlaxAlbertModel.from_pretrained("albert-base-v2" ) _UpperCAmelCase : List[Any] = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) _UpperCAmelCase : Optional[int] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCAmelCase : Dict = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )[0] _UpperCAmelCase : List[Any] = (1, 11, 7_68) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : str = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase__ , atol=1E-4 ) )
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Tuple = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def __lowerCAmelCase (): print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
357
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-classification/requirements.txt') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "rb" ) as f: _UpperCAmelCase : List[str] = Image.open(__lowerCAmelCase ) return im.convert("RGB" ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={ "help": "Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub)." } , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the training data."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the validation data."} ) lowerCAmelCase : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default="google/vit-base-patch16-224-in21k" , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCAmelCase__ )} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) lowerCAmelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCAmelCase : str = field(default=UpperCAmelCase__ , metadata={"help": "Name or path of preprocessor config."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = torch.stack([example["pixel_values"] for example in examples] ) _UpperCAmelCase : Tuple = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def __lowerCAmelCase (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , __lowerCAmelCase , __lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(__lowerCAmelCase ) transformers.utils.logging.set_verbosity(__lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _UpperCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCAmelCase : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCAmelCase : List[Any] = {} if data_args.train_dir is not None: _UpperCAmelCase : str = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCAmelCase : Optional[Any] = os.path.join(data_args.validation_dir , "**" ) _UpperCAmelCase : Any = load_dataset( "imagefolder" , data_files=__lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCAmelCase : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __lowerCAmelCase ) and data_args.train_val_split > 0.0: _UpperCAmelCase : List[Any] = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCAmelCase : List[str] = split["train"] _UpperCAmelCase : Union[str, Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase : Optional[int] = dataset["train"].features["labels"].names _UpperCAmelCase , _UpperCAmelCase : int = {}, {} for i, label in enumerate(__lowerCAmelCase ): _UpperCAmelCase : int = str(__lowerCAmelCase ) _UpperCAmelCase : str = label # Load the accuracy metric from the datasets package _UpperCAmelCase : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(__lowerCAmelCase ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowerCAmelCase ) , labelaid=__lowerCAmelCase , idalabel=__lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCAmelCase : Dict = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCAmelCase : int = image_processor.size["shortest_edge"] else: _UpperCAmelCase : int = (image_processor.size["height"], image_processor.size["width"]) _UpperCAmelCase : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCAmelCase : Optional[int] = Compose( [ RandomResizedCrop(__lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCAmelCase : Union[str, Any] = Compose( [ Resize(__lowerCAmelCase ), CenterCrop(__lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCAmelCase : Dict = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(__lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCAmelCase : Optional[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(__lowerCAmelCase ) # Initalize our trainer _UpperCAmelCase : Union[str, Any] = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=__lowerCAmelCase , tokenizer=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : Any = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : int = last_checkpoint _UpperCAmelCase : Dict = trainer.train(resume_from_checkpoint=__lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowerCAmelCase ) trainer.save_metrics("eval" , __lowerCAmelCase ) # Write model card and (optionally) push to hub _UpperCAmelCase : int = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowerCAmelCase ) else: trainer.create_model_card(**__lowerCAmelCase ) if __name__ == "__main__": main()
322
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : List[Any] = "mvp" lowerCAmelCase : int = ["past_key_values"] lowerCAmelCase : Union[str, Any] = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : List[Any] , lowerCamelCase__ : int=5_02_67 , lowerCamelCase__ : str=10_24 , lowerCamelCase__ : Optional[Any]=12 , lowerCamelCase__ : Optional[int]=40_96 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : Union[str, Any]=12 , lowerCamelCase__ : Dict=40_96 , lowerCamelCase__ : List[Any]=16 , lowerCamelCase__ : Tuple=0.0 , lowerCamelCase__ : int=0.0 , lowerCamelCase__ : int="gelu" , lowerCamelCase__ : Optional[int]=10_24 , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Any=0.0 , lowerCamelCase__ : Dict=0.0 , lowerCamelCase__ : str=0.0_2 , lowerCamelCase__ : Union[str, Any]=0.0 , lowerCamelCase__ : List[str]=False , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : int=1 , lowerCamelCase__ : Optional[int]=0 , lowerCamelCase__ : Optional[int]=2 , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : int=2 , lowerCamelCase__ : Optional[int]=2 , lowerCamelCase__ : Dict=False , lowerCamelCase__ : str=1_00 , lowerCamelCase__ : List[Any]=8_00 , **lowerCamelCase__ : Optional[int] , ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = vocab_size _UpperCAmelCase : Any = max_position_embeddings _UpperCAmelCase : Dict = d_model _UpperCAmelCase : Union[str, Any] = encoder_ffn_dim _UpperCAmelCase : Union[str, Any] = encoder_layers _UpperCAmelCase : str = encoder_attention_heads _UpperCAmelCase : Tuple = decoder_ffn_dim _UpperCAmelCase : str = decoder_layers _UpperCAmelCase : List[Any] = decoder_attention_heads _UpperCAmelCase : Optional[int] = dropout _UpperCAmelCase : Dict = attention_dropout _UpperCAmelCase : Optional[int] = activation_dropout _UpperCAmelCase : List[Any] = activation_function _UpperCAmelCase : Union[str, Any] = init_std _UpperCAmelCase : Dict = encoder_layerdrop _UpperCAmelCase : List[Any] = decoder_layerdrop _UpperCAmelCase : List[Any] = classifier_dropout _UpperCAmelCase : str = use_cache _UpperCAmelCase : str = encoder_layers _UpperCAmelCase : List[str] = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCAmelCase : Tuple = use_prompt _UpperCAmelCase : List[str] = prompt_length _UpperCAmelCase : Any = prompt_mid_dim super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , is_encoder_decoder=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , forced_eos_token_id=lowerCamelCase__ , **lowerCamelCase__ , ) if self.forced_bos_token_id is None and kwargs.get("force_bos_token_to_be_generated" , lowerCamelCase__ ): _UpperCAmelCase : Optional[int] = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ "The config can simply be saved and uploaded again to be fixed." )
358
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase__ = logging.get_logger(__name__) # General docstring lowerCamelCase__ = 'RegNetConfig' # Base docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = [1, 1_088, 7, 7] # Image classification docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = 'tabby, tabby cat' lowerCamelCase__ = [ 'facebook/regnet-y-040', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 3 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[str] = "relu" , **lowerCamelCase__ : Tuple , ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _UpperCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _UpperCAmelCase : Dict = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=lowerCamelCase__ , strides=lowerCamelCase__ , padding="VALID" , groups=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" , ) _UpperCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) _UpperCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.convolution(self.padding(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = self.normalization(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = config.num_channels _UpperCAmelCase : Any = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = shape_list(lowerCamelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _UpperCAmelCase : Optional[Any] = tf.transpose(lowerCamelCase__ , perm=(0, 2, 3, 1) ) _UpperCAmelCase : List[Any] = self.embedder(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : int = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=1 , strides=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" ) _UpperCAmelCase : Any = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False ) ->tf.Tensor: '''simple docstring''' return self.normalization(self.convolution(lowerCamelCase__ ) , training=lowerCamelCase__ ) class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : int , **lowerCamelCase__ : Optional[int] ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) _UpperCAmelCase : int = [ tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.pooler(lowerCamelCase__ ) for layer_module in self.attention: _UpperCAmelCase : str = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : Any ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = in_channels != out_channels or stride != 1 _UpperCAmelCase : List[str] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : List[str] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _UpperCAmelCase : Optional[int] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.2" ), ] _UpperCAmelCase : Union[str, Any] = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = hidden_state for layer_module in self.layers: _UpperCAmelCase : List[Any] = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : List[Any] , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Optional[int] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : Union[str, Any] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) _UpperCAmelCase : List[Any] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(lowerCamelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.3" ), ] _UpperCAmelCase : int = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str ) ->Any: '''simple docstring''' _UpperCAmelCase : int = hidden_state for layer_module in self.layers: _UpperCAmelCase : Tuple = layer_module(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : Tuple = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : str = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer _UpperCAmelCase : List[str] = [ # downsampling is done in the first layer with stride of 2 layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , name="layers.0" ), *[layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , name=F"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] ) ->List[str]: '''simple docstring''' for layer_module in self.layers: _UpperCAmelCase : Optional[int] = layer_module(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : int ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) _UpperCAmelCase : Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCamelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , depth=lowerCamelCase__ , name=F"""stages.{i+1}""" ) ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = True ) ->TFBaseModelOutputWithNoAttention: '''simple docstring''' _UpperCAmelCase : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase : Optional[Any] = hidden_states + (hidden_state,) _UpperCAmelCase : Dict = stage_module(lowerCamelCase__ ) if output_hidden_states: _UpperCAmelCase : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase__ , hidden_states=lowerCamelCase__ ) @keras_serializable class lowerCAmelCase__ ( tf.keras.layers.Layer ): lowerCAmelCase : Optional[Any] = RegNetConfig def __init__( self : Union[str, Any] , lowerCamelCase__ : Any , **lowerCamelCase__ : str ) ->int: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = config _UpperCAmelCase : Union[str, Any] = TFRegNetEmbeddings(lowerCamelCase__ , name="embedder" ) _UpperCAmelCase : Union[str, Any] = TFRegNetEncoder(lowerCamelCase__ , name="encoder" ) _UpperCAmelCase : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) @unpack_inputs def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : bool = False , ) ->TFBaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' _UpperCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.embedder(lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : str = self.encoder( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : Dict = encoder_outputs[0] _UpperCAmelCase : Dict = self.pooler(lowerCamelCase__ ) # Change to NCHW output format have uniformity in the modules _UpperCAmelCase : Union[str, Any] = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) _UpperCAmelCase : Tuple = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _UpperCAmelCase : List[str] = tuple([tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase__ , pooler_output=lowerCamelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = RegNetConfig lowerCAmelCase : Tuple = "regnet" lowerCAmelCase : Union[str, Any] = "pixel_values" @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} lowerCamelCase__ = r'\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCamelCase__ = r'\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : Any , **lowerCamelCase__ : List[str] ) ->Optional[int]: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Any=False , ) ->Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( pixel_values=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : List[Any] , **lowerCamelCase__ : Union[str, Any] ) ->Any: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = config.num_labels _UpperCAmelCase : Dict = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) # classification head _UpperCAmelCase : str = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict=False , ) ->Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : str = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : int = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase : Dict = self.classifier[0](lowerCamelCase__ ) _UpperCAmelCase : str = self.classifier[1](lowerCamelCase__ ) _UpperCAmelCase : Tuple = None if labels is None else self.hf_compute_loss(labels=lowerCamelCase__ , logits=lowerCamelCase__ ) if not return_dict: _UpperCAmelCase : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCamelCase__ , logits=lowerCamelCase__ , hidden_states=outputs.hidden_states )
322
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} lowerCamelCase__ = { 'vocab_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json' ), }, } lowerCamelCase__ = { 'yjernite/retribert-base-uncased': 512, } lowerCamelCase__ = { 'yjernite/retribert-base-uncased': {'do_lower_case': True}, } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Dict = VOCAB_FILES_NAMES lowerCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase : List[Any] = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase : int = RetriBertTokenizer lowerCAmelCase : int = ["input_ids", "attention_mask"] def __init__( self : int , lowerCamelCase__ : List[str]=None , lowerCamelCase__ : Dict=None , lowerCamelCase__ : Union[str, Any]=True , lowerCamelCase__ : List[Any]="[UNK]" , lowerCamelCase__ : List[Any]="[SEP]" , lowerCamelCase__ : Dict="[PAD]" , lowerCamelCase__ : List[str]="[CLS]" , lowerCamelCase__ : Tuple="[MASK]" , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : List[str]=None , **lowerCamelCase__ : int , ) ->int: '''simple docstring''' super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) _UpperCAmelCase : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCamelCase__ ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCamelCase__ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCamelCase__ ) != tokenize_chinese_chars ): _UpperCAmelCase : Dict = getattr(lowerCamelCase__ , normalizer_state.pop("type" ) ) _UpperCAmelCase : Optional[int] = do_lower_case _UpperCAmelCase : Optional[Any] = strip_accents _UpperCAmelCase : str = tokenize_chinese_chars _UpperCAmelCase : Optional[int] = normalizer_class(**lowerCamelCase__ ) _UpperCAmelCase : List[Any] = do_lower_case def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Tuple=None ) ->Tuple: '''simple docstring''' _UpperCAmelCase : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : List[int] , lowerCamelCase__ : Optional[List[int]] = None ) ->List[int]: '''simple docstring''' _UpperCAmelCase : int = [self.sep_token_id] _UpperCAmelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str , lowerCamelCase__ : Optional[str] = None ) ->Tuple[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
359
'''simple docstring''' import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def __lowerCAmelCase (__lowerCAmelCase ): if is_torch_version("<" , "2.0.0" ) or not hasattr(__lowerCAmelCase , "_dynamo" ): return False return isinstance(__lowerCAmelCase , torch._dynamo.eval_frame.OptimizedModule ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase = True ): _UpperCAmelCase : Any = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) _UpperCAmelCase : Dict = is_compiled_module(__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : Optional[int] = model _UpperCAmelCase : Any = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = model.module if not keep_fpaa_wrapper: _UpperCAmelCase : List[Any] = getattr(__lowerCAmelCase , "forward" ) _UpperCAmelCase : Dict = model.__dict__.pop("_original_forward" , __lowerCAmelCase ) if original_forward is not None: while hasattr(__lowerCAmelCase , "__wrapped__" ): _UpperCAmelCase : Optional[int] = forward.__wrapped__ if forward == original_forward: break _UpperCAmelCase : Dict = forward if getattr(__lowerCAmelCase , "_converted_to_transformer_engine" , __lowerCAmelCase ): convert_model(__lowerCAmelCase , to_transformer_engine=__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : int = model _UpperCAmelCase : str = compiled_model return model def __lowerCAmelCase (): PartialState().wait_for_everyone() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(__lowerCAmelCase , __lowerCAmelCase ) elif PartialState().local_process_index == 0: torch.save(__lowerCAmelCase , __lowerCAmelCase ) @contextmanager def __lowerCAmelCase (**__lowerCAmelCase ): for key, value in kwargs.items(): _UpperCAmelCase : str = str(__lowerCAmelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def __lowerCAmelCase (__lowerCAmelCase ): if not hasattr(__lowerCAmelCase , "__qualname__" ) and not hasattr(__lowerCAmelCase , "__name__" ): _UpperCAmelCase : List[str] = getattr(__lowerCAmelCase , "__class__" , __lowerCAmelCase ) if hasattr(__lowerCAmelCase , "__qualname__" ): return obj.__qualname__ if hasattr(__lowerCAmelCase , "__name__" ): return obj.__name__ return str(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): for key, value in source.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = destination.setdefault(__lowerCAmelCase , {} ) merge_dicts(__lowerCAmelCase , __lowerCAmelCase ) else: _UpperCAmelCase : Optional[int] = value return destination def __lowerCAmelCase (__lowerCAmelCase = None ): if port is None: _UpperCAmelCase : Tuple = 29_500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
322
0
'''simple docstring''' import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency lowerCamelCase__ = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } lowerCamelCase__ = 'ETAOINSHRDLCUMWFGYPBVKJXQZ' lowerCamelCase__ = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def __lowerCAmelCase (__lowerCAmelCase ): return x[0] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : int = get_letter_count(__lowerCAmelCase ) _UpperCAmelCase : dict[int, list[str]] = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(__lowerCAmelCase ) _UpperCAmelCase : dict[int, str] = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = "".join(freq_to_letter[freq] ) _UpperCAmelCase : Any = list(freq_to_letter_str.items() ) freq_pairs.sort(key=__lowerCAmelCase , reverse=__lowerCAmelCase ) _UpperCAmelCase : list[str] = [freq_pair[1] for freq_pair in freq_pairs] return "".join(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Any = get_frequency_order(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
360
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): if number < 0: raise ValueError("number must not be negative" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
322
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowerCamelCase__ = {'configuration_deit': ['DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DeiTConfig', 'DeiTOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ['DeiTFeatureExtractor'] lowerCamelCase__ = ['DeiTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ 'DEIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'DeiTForImageClassification', 'DeiTForImageClassificationWithTeacher', 'DeiTForMaskedImageModeling', 'DeiTModel', 'DeiTPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ 'TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDeiTForImageClassification', 'TFDeiTForImageClassificationWithTeacher', 'TFDeiTForMaskedImageModeling', 'TFDeiTModel', 'TFDeiTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_deit import DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, DeiTConfig, DeiTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_deit import DeiTFeatureExtractor from .image_processing_deit import DeiTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deit import ( DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, DeiTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deit import ( TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, TFDeiTPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
361
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('Program to check whether a number is a Perfect number or not...') lowerCamelCase__ = int(input('Enter number: ').strip()) print(F'''{number} is {"" if perfect(number) else "not "}a Perfect Number.''')
322
0
'''simple docstring''' from __future__ import annotations lowerCamelCase__ = 10 def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Tuple = 1 _UpperCAmelCase : Dict = max(__lowerCAmelCase ) while placement <= max_digit: # declare and initialize empty buckets _UpperCAmelCase : list[list] = [[] for _ in range(__lowerCAmelCase )] # split list_of_ints between the buckets for i in list_of_ints: _UpperCAmelCase : Optional[int] = int((i / placement) % RADIX ) buckets[tmp].append(__lowerCAmelCase ) # put each buckets' contents into list_of_ints _UpperCAmelCase : List[str] = 0 for b in range(__lowerCAmelCase ): for i in buckets[b]: _UpperCAmelCase : str = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
362
'''simple docstring''' from collections.abc import Sequence def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return sum(c * (x**i) for i, c in enumerate(__lowerCAmelCase ) ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Dict = 0.0 for coeff in reversed(__lowerCAmelCase ): _UpperCAmelCase : int = result * x + coeff return result if __name__ == "__main__": lowerCamelCase__ = (0.0, 0.0, 5.0, 9.3, 7.0) lowerCamelCase__ = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
322
0
'''simple docstring''' import cva import numpy as np class lowerCAmelCase__ : def __init__( self : str , lowerCamelCase__ : float , lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' if k in (0.0_4, 0.0_6): _UpperCAmelCase : Optional[Any] = k _UpperCAmelCase : Dict = window_size else: raise ValueError("invalid k value" ) def __str__( self : Optional[Any] ) ->str: '''simple docstring''' return str(self.k ) def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : str ) ->tuple[cva.Mat, list[list[int]]]: '''simple docstring''' _UpperCAmelCase : int = cva.imread(lowerCamelCase__ , 0 ) _UpperCAmelCase : Dict = img.shape _UpperCAmelCase : list[list[int]] = [] _UpperCAmelCase : Any = img.copy() _UpperCAmelCase : int = cva.cvtColor(lowerCamelCase__ , cva.COLOR_GRAY2RGB ) _UpperCAmelCase : int = np.gradient(lowerCamelCase__ ) _UpperCAmelCase : Tuple = dx**2 _UpperCAmelCase : List[str] = dy**2 _UpperCAmelCase : Optional[int] = dx * dy _UpperCAmelCase : Union[str, Any] = 0.0_4 _UpperCAmelCase : Optional[int] = self.window_size // 2 for y in range(lowerCamelCase__ , h - offset ): for x in range(lowerCamelCase__ , w - offset ): _UpperCAmelCase : List[Any] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _UpperCAmelCase : str = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _UpperCAmelCase : Union[str, Any] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _UpperCAmelCase : List[Any] = (wxx * wyy) - (wxy**2) _UpperCAmelCase : List[Any] = wxx + wyy _UpperCAmelCase : Dict = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 2_55 ) return color_img, corner_list if __name__ == "__main__": lowerCamelCase__ = HarrisCorner(0.04, 3) lowerCamelCase__ ,lowerCamelCase__ = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
363
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[Any] = len(__lowerCAmelCase ) _UpperCAmelCase : Tuple = sum(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _UpperCAmelCase : Any = True for i in range(1 , s + 1 ): _UpperCAmelCase : List[Any] = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): _UpperCAmelCase : Optional[int] = dp[i][j - 1] if arr[i - 1] <= j: _UpperCAmelCase : Any = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: _UpperCAmelCase : List[Any] = s - 2 * j break return diff
322
0
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = ["image_processor", "tokenizer"] lowerCAmelCase : Tuple = "LayoutLMv3ImageProcessor" lowerCAmelCase : Tuple = ("LayoutLMv3Tokenizer", "LayoutLMv3TokenizerFast") def __init__( self : Dict , lowerCamelCase__ : Union[str, Any]=None , lowerCamelCase__ : Optional[Any]=None , **lowerCamelCase__ : Any ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Optional[int] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowerCamelCase__ , ) _UpperCAmelCase : Any = kwargs.pop("feature_extractor" ) _UpperCAmelCase : List[str] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowerCamelCase__ , lowerCamelCase__ ) def __call__( self : Any , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , lowerCamelCase__ : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , lowerCamelCase__ : Union[List[List[int]], List[List[List[int]]]] = None , lowerCamelCase__ : Optional[Union[List[int], List[List[int]]]] = None , lowerCamelCase__ : bool = True , lowerCamelCase__ : Union[bool, str, PaddingStrategy] = False , lowerCamelCase__ : Union[bool, str, TruncationStrategy] = None , lowerCamelCase__ : Optional[int] = None , lowerCamelCase__ : int = 0 , lowerCamelCase__ : Optional[int] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = True , lowerCamelCase__ : Optional[Union[str, TensorType]] = None , **lowerCamelCase__ : Tuple , ) ->BatchEncoding: '''simple docstring''' if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( "You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True." ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( "You cannot provide word labels if you initialized the image processor with apply_ocr set to True." ) # first, apply the image processor _UpperCAmelCase : Any = self.image_processor(images=lowerCamelCase__ , return_tensors=lowerCamelCase__ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Union[str, Any] = [text] # add batch dimension (as the image processor always adds a batch dimension) _UpperCAmelCase : List[str] = features["words"] _UpperCAmelCase : int = self.tokenizer( text=text if text is not None else features["words"] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["boxes"] , word_labels=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , max_length=lowerCamelCase__ , stride=lowerCamelCase__ , pad_to_multiple_of=lowerCamelCase__ , return_token_type_ids=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , return_overflowing_tokens=lowerCamelCase__ , return_special_tokens_mask=lowerCamelCase__ , return_offsets_mapping=lowerCamelCase__ , return_length=lowerCamelCase__ , verbose=lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ , ) # add pixel values _UpperCAmelCase : int = features.pop("pixel_values" ) if return_overflowing_tokens is True: _UpperCAmelCase : str = self.get_overflowing_images(lowerCamelCase__ , encoded_inputs["overflow_to_sample_mapping"] ) _UpperCAmelCase : Any = images return encoded_inputs def lowerCAmelCase__ ( self : int , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(lowerCamelCase__ ) != len(lowerCamelCase__ ): raise ValueError( "Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got" F""" {len(lowerCamelCase__ )} and {len(lowerCamelCase__ )}""" ) return images_with_overflow def lowerCAmelCase__ ( self : Union[str, Any] , *lowerCamelCase__ : int , **lowerCamelCase__ : Optional[Any] ) ->str: '''simple docstring''' return self.tokenizer.batch_decode(*lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] , *lowerCamelCase__ : List[str] , **lowerCamelCase__ : Dict ) ->Tuple: '''simple docstring''' return self.tokenizer.decode(*lowerCamelCase__ , **lowerCamelCase__ ) @property def lowerCAmelCase__ ( self : Tuple ) ->List[str]: '''simple docstring''' return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def lowerCAmelCase__ ( self : Union[str, Any] ) ->List[Any]: '''simple docstring''' warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , lowerCamelCase__ , ) return self.image_processor_class @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Dict: '''simple docstring''' warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , lowerCamelCase__ , ) return self.image_processor
364
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'microsoft/resnet-50': 'https://huggingface.co/microsoft/resnet-50/blob/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): lowerCAmelCase : int = "resnet" lowerCAmelCase : Union[str, Any] = ["basic", "bottleneck"] def __init__( self : Dict , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : Any=64 , lowerCamelCase__ : Optional[int]=[2_56, 5_12, 10_24, 20_48] , lowerCamelCase__ : int=[3, 4, 6, 3] , lowerCamelCase__ : Dict="bottleneck" , lowerCamelCase__ : Dict="relu" , lowerCamelCase__ : List[Any]=False , lowerCamelCase__ : Any=None , lowerCamelCase__ : int=None , **lowerCamelCase__ : Tuple , ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) _UpperCAmelCase : str = num_channels _UpperCAmelCase : List[str] = embedding_size _UpperCAmelCase : Tuple = hidden_sizes _UpperCAmelCase : Dict = depths _UpperCAmelCase : List[Any] = layer_type _UpperCAmelCase : Optional[int] = hidden_act _UpperCAmelCase : Tuple = downsample_in_first_stage _UpperCAmelCase : str = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(lowerCamelCase__ ) + 1 )] _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = get_aligned_output_features_output_indices( out_features=lowerCamelCase__ , out_indices=lowerCamelCase__ , stage_names=self.stage_names ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = version.parse("1.11" ) @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowerCAmelCase__ ( self : str ) ->float: '''simple docstring''' return 1E-3
322
0
'''simple docstring''' import os import sys import transformers lowerCamelCase__ = '3' print('Python version:', sys.version) print('transformers version:', transformers.__version__) try: import torch print('Torch version:', torch.__version__) print('Cuda available:', torch.cuda.is_available()) print('Cuda version:', torch.version.cuda) print('CuDNN version:', torch.backends.cudnn.version()) print('Number of GPUs available:', torch.cuda.device_count()) print('NCCL version:', torch.cuda.nccl.version()) except ImportError: print('Torch version:', None) try: import deepspeed print('DeepSpeed version:', deepspeed.__version__) except ImportError: print('DeepSpeed version:', None) try: import tensorflow as tf print('TensorFlow version:', tf.__version__) print('TF GPUs available:', bool(tf.config.list_physical_devices('GPU'))) print('Number of TF GPUs available:', len(tf.config.list_physical_devices('GPU'))) except ImportError: print('TensorFlow version:', None)
365
'''simple docstring''' from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowerCamelCase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __lowerCAmelCase (__lowerCAmelCase ): if isinstance(__lowerCAmelCase , torch.Tensor ): return image elif isinstance(__lowerCAmelCase , PIL.Image.Image ): _UpperCAmelCase : int = [image] _UpperCAmelCase : str = [trans(img.convert("RGB" ) ) for img in image] _UpperCAmelCase : Optional[Any] = torch.stack(__lowerCAmelCase ) return image class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Tuple , lowerCamelCase__ : int , lowerCamelCase__ : int ) ->int: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM _UpperCAmelCase : Tuple = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : str ) ->Union[str, Any]: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(F"""The value of strength should in [0.0, 1.0] but is {strength}""" ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Dict , lowerCamelCase__ : List[str] , lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = min(int(num_inference_steps * strength ) , lowerCamelCase__ ) _UpperCAmelCase : str = max(num_inference_steps - init_timestep , 0 ) _UpperCAmelCase : List[str] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : str , lowerCamelCase__ : str , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[Any]=None ) ->str: '''simple docstring''' if not isinstance(lowerCamelCase__ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowerCamelCase__ )}""" ) _UpperCAmelCase : Union[str, Any] = image.to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(lowerCamelCase__ )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) _UpperCAmelCase : List[str] = init_latents.shape _UpperCAmelCase : Optional[int] = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=lowerCamelCase__ , dtype=lowerCamelCase__ ) # get latents print("add noise to latents at timestep" , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.scheduler.add_noise(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = init_latents return latents @torch.no_grad() def __call__( self : Any , lowerCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image] = None , lowerCamelCase__ : float = 0.8 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase__ : float = 0.0 , lowerCamelCase__ : int = 50 , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , ) ->Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowerCamelCase__ ) # 2. Preprocess image _UpperCAmelCase : Dict = preprocess(lowerCamelCase__ ) # 3. set timesteps self.scheduler.set_timesteps(lowerCamelCase__ , device=self.device ) _UpperCAmelCase , _UpperCAmelCase : Any = self.get_timesteps(lowerCamelCase__ , lowerCamelCase__ , self.device ) _UpperCAmelCase : List[Any] = timesteps[:1].repeat(lowerCamelCase__ ) # 4. Prepare latent variables _UpperCAmelCase : Optional[int] = self.prepare_latents(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , self.unet.dtype , self.device , lowerCamelCase__ ) _UpperCAmelCase : Any = latents # 5. Denoising loop for t in self.progress_bar(lowerCamelCase__ ): # 1. predict noise model_output _UpperCAmelCase : Union[str, Any] = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _UpperCAmelCase : int = self.scheduler.step( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , eta=lowerCamelCase__ , use_clipped_model_output=lowerCamelCase__ , generator=lowerCamelCase__ , ).prev_sample _UpperCAmelCase : Dict = (image / 2 + 0.5).clamp(0 , 1 ) _UpperCAmelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCAmelCase : str = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowerCamelCase__ )
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ): raise ValueError("iterations must be defined as integers" ) if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not number >= 1: raise ValueError( "starting number must be\n and integer and be more than 0" ) if not iterations >= 1: raise ValueError("Iterations must be done more than 0 times to play FizzBuzz" ) _UpperCAmelCase : Tuple = "" while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(__lowerCAmelCase ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
366
'''simple docstring''' from __future__ import annotations from collections.abc import Callable lowerCamelCase__ = list[list[float | int]] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(size + 1 )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : float for row in range(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = matrix[row][col] _UpperCAmelCase : Optional[int] = vector[row][0] _UpperCAmelCase : int = 0 _UpperCAmelCase : Union[str, Any] = 0 while row < size and col < size: # pivoting _UpperCAmelCase : Optional[Any] = max((abs(augmented[rowa][col] ), rowa) for rowa in range(__lowerCAmelCase , __lowerCAmelCase ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: _UpperCAmelCase , _UpperCAmelCase : str = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , __lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[rowa][col] / augmented[row][col] _UpperCAmelCase : Optional[Any] = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , __lowerCAmelCase ): for row in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[row][col] / augmented[col][col] for cola in range(__lowerCAmelCase , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(__lowerCAmelCase ) ] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(__lowerCAmelCase )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix = [[0] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int for x_val, y_val in enumerate(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = (x_val + 1) ** (size - col - 1) _UpperCAmelCase : int = y_val _UpperCAmelCase : List[str] = solve(__lowerCAmelCase , __lowerCAmelCase ) def interpolated_func(__lowerCAmelCase ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(__lowerCAmelCase ) ) return interpolated_func def __lowerCAmelCase (__lowerCAmelCase ): return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def __lowerCAmelCase (__lowerCAmelCase = question_function , __lowerCAmelCase = 10 ): _UpperCAmelCase : list[int] = [func(__lowerCAmelCase ) for x_val in range(1 , order + 1 )] _UpperCAmelCase : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] _UpperCAmelCase : int = 0 _UpperCAmelCase : Callable[[int], int] _UpperCAmelCase : int for poly in polynomials: _UpperCAmelCase : int = 1 while func(__lowerCAmelCase ) == poly(__lowerCAmelCase ): x_val += 1 ret += poly(__lowerCAmelCase ) return ret if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline lowerCamelCase__ = version.parse(version.parse(torch.__version__).base_version) < version.parse('1.11') def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False , ): output_path.parent.mkdir(parents=__lowerCAmelCase , exist_ok=__lowerCAmelCase ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( __lowerCAmelCase , __lowerCAmelCase , f=output_path.as_posix() , input_names=__lowerCAmelCase , output_names=__lowerCAmelCase , dynamic_axes=__lowerCAmelCase , do_constant_folding=__lowerCAmelCase , use_external_data_format=__lowerCAmelCase , enable_onnx_checker=__lowerCAmelCase , opset_version=__lowerCAmelCase , ) else: export( __lowerCAmelCase , __lowerCAmelCase , f=output_path.as_posix() , input_names=__lowerCAmelCase , output_names=__lowerCAmelCase , dynamic_axes=__lowerCAmelCase , do_constant_folding=__lowerCAmelCase , opset_version=__lowerCAmelCase , ) @torch.no_grad() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = False ): _UpperCAmelCase : int = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): _UpperCAmelCase : Optional[int] = "cuda" elif fpaa and not torch.cuda.is_available(): raise ValueError("`float16` model export is only supported on GPUs with CUDA" ) else: _UpperCAmelCase : Tuple = "cpu" _UpperCAmelCase : Any = StableDiffusionPipeline.from_pretrained(__lowerCAmelCase , torch_dtype=__lowerCAmelCase ).to(__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = Path(__lowerCAmelCase ) # TEXT ENCODER _UpperCAmelCase : Union[str, Any] = pipeline.text_encoder.config.max_position_embeddings _UpperCAmelCase : Optional[int] = pipeline.text_encoder.config.hidden_size _UpperCAmelCase : Optional[int] = pipeline.tokenizer( "A sample prompt" , padding="max_length" , max_length=pipeline.tokenizer.model_max_length , truncation=__lowerCAmelCase , return_tensors="pt" , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=__lowerCAmelCase , dtype=torch.intaa )) , output_path=output_path / "text_encoder" / "model.onnx" , ordered_input_names=["input_ids"] , output_names=["last_hidden_state", "pooler_output"] , dynamic_axes={ "input_ids": {0: "batch", 1: "sequence"}, } , opset=__lowerCAmelCase , ) del pipeline.text_encoder # UNET _UpperCAmelCase : Optional[int] = pipeline.unet.config.in_channels _UpperCAmelCase : Any = pipeline.unet.config.sample_size _UpperCAmelCase : Optional[int] = output_path / "unet" / "model.onnx" onnx_export( pipeline.unet , model_args=( torch.randn(2 , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ).to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ), torch.randn(2 ).to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ), torch.randn(2 , __lowerCAmelCase , __lowerCAmelCase ).to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ), False, ) , output_path=__lowerCAmelCase , ordered_input_names=["sample", "timestep", "encoder_hidden_states", "return_dict"] , output_names=["out_sample"] , dynamic_axes={ "sample": {0: "batch", 1: "channels", 2: "height", 3: "width"}, "timestep": {0: "batch"}, "encoder_hidden_states": {0: "batch", 1: "sequence"}, } , opset=__lowerCAmelCase , use_external_data_format=__lowerCAmelCase , ) _UpperCAmelCase : List[str] = str(unet_path.absolute().as_posix() ) _UpperCAmelCase : int = os.path.dirname(__lowerCAmelCase ) _UpperCAmelCase : Any = onnx.load(__lowerCAmelCase ) # clean up existing tensor files shutil.rmtree(__lowerCAmelCase ) os.mkdir(__lowerCAmelCase ) # collate external tensor files into one onnx.save_model( __lowerCAmelCase , __lowerCAmelCase , save_as_external_data=__lowerCAmelCase , all_tensors_to_one_file=__lowerCAmelCase , location="weights.pb" , convert_attribute=__lowerCAmelCase , ) del pipeline.unet # VAE ENCODER _UpperCAmelCase : List[Any] = pipeline.vae _UpperCAmelCase : Tuple = vae_encoder.config.in_channels _UpperCAmelCase : List[Any] = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder _UpperCAmelCase : Tuple = lambda __lowerCAmelCase , __lowerCAmelCase : vae_encoder.encode(__lowerCAmelCase , __lowerCAmelCase )[0].sample() onnx_export( __lowerCAmelCase , model_args=( torch.randn(1 , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ).to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ), False, ) , output_path=output_path / "vae_encoder" / "model.onnx" , ordered_input_names=["sample", "return_dict"] , output_names=["latent_sample"] , dynamic_axes={ "sample": {0: "batch", 1: "channels", 2: "height", 3: "width"}, } , opset=__lowerCAmelCase , ) # VAE DECODER _UpperCAmelCase : List[str] = pipeline.vae _UpperCAmelCase : Dict = vae_decoder.config.latent_channels _UpperCAmelCase : int = vae_decoder.config.out_channels # forward only through the decoder part _UpperCAmelCase : str = vae_encoder.decode onnx_export( __lowerCAmelCase , model_args=( torch.randn(1 , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ).to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ), False, ) , output_path=output_path / "vae_decoder" / "model.onnx" , ordered_input_names=["latent_sample", "return_dict"] , output_names=["sample"] , dynamic_axes={ "latent_sample": {0: "batch", 1: "channels", 2: "height", 3: "width"}, } , opset=__lowerCAmelCase , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: _UpperCAmelCase : Optional[Any] = pipeline.safety_checker _UpperCAmelCase : List[Any] = safety_checker.config.vision_config.num_channels _UpperCAmelCase : List[str] = safety_checker.config.vision_config.image_size _UpperCAmelCase : Union[str, Any] = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ).to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ), torch.randn(1 , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ).to(device=__lowerCAmelCase , dtype=__lowerCAmelCase ), ) , output_path=output_path / "safety_checker" / "model.onnx" , ordered_input_names=["clip_input", "images"] , output_names=["out_images", "has_nsfw_concepts"] , dynamic_axes={ "clip_input": {0: "batch", 1: "channels", 2: "height", 3: "width"}, "images": {0: "batch", 1: "height", 2: "width", 3: "channels"}, } , opset=__lowerCAmelCase , ) del pipeline.safety_checker _UpperCAmelCase : Dict = OnnxRuntimeModel.from_pretrained(output_path / "safety_checker" ) _UpperCAmelCase : List[str] = pipeline.feature_extractor else: _UpperCAmelCase : List[str] = None _UpperCAmelCase : int = None _UpperCAmelCase : List[str] = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / "vae_encoder" ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / "vae_decoder" ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / "text_encoder" ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / "unet" ) , scheduler=pipeline.scheduler , safety_checker=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(__lowerCAmelCase ) print("ONNX pipeline saved to" , __lowerCAmelCase ) del pipeline del onnx_pipeline _UpperCAmelCase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(__lowerCAmelCase , provider="CPUExecutionProvider" ) print("ONNX pipeline is loadable" ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument( '--model_path', type=str, required=True, help='Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).', ) parser.add_argument('--output_path', type=str, required=True, help='Path to the output model.') parser.add_argument( '--opset', default=14, type=int, help='The version of the ONNX operator set to use.', ) parser.add_argument('--fp16', action='store_true', default=False, help='Export the models in `float16` mode') lowerCamelCase__ = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
367
'''simple docstring''' from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
322
0
'''simple docstring''' from __future__ import annotations lowerCamelCase__ = 'Muhammad Umer Farooq' lowerCamelCase__ = 'MIT' lowerCamelCase__ = '1.0.0' lowerCamelCase__ = 'Muhammad Umer Farooq' lowerCamelCase__ = 'contact@muhammadumerfarooq.me' lowerCamelCase__ = 'Alpha' import re from html.parser import HTMLParser from urllib import parse import requests class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Optional[int] , lowerCamelCase__ : str ) ->None: '''simple docstring''' super().__init__() _UpperCAmelCase : list[str] = [] _UpperCAmelCase : List[str] = domain def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : str , lowerCamelCase__ : list[tuple[str, str | None]] ) ->None: '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: _UpperCAmelCase : List[Any] = parse.urljoin(self.domain , lowerCamelCase__ ) self.urls.append(lowerCamelCase__ ) def __lowerCAmelCase (__lowerCAmelCase ): return ".".join(get_sub_domain_name(__lowerCAmelCase ).split("." )[-2:] ) def __lowerCAmelCase (__lowerCAmelCase ): return parse.urlparse(__lowerCAmelCase ).netloc def __lowerCAmelCase (__lowerCAmelCase = "https://github.com" ): _UpperCAmelCase : str = get_domain_name(__lowerCAmelCase ) # Initialize the parser _UpperCAmelCase : Any = Parser(__lowerCAmelCase ) try: # Open URL _UpperCAmelCase : List[Any] = requests.get(__lowerCAmelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through _UpperCAmelCase : Optional[Any] = set() for link in parser.urls: # open URL. # read = requests.get(link) try: _UpperCAmelCase : Any = requests.get(__lowerCAmelCase ) # Get the valid email. _UpperCAmelCase : List[Any] = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(__lowerCAmelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(__lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase__ = emails_from_url('https://github.com') print(F'''{len(emails)} emails found:''') print('\n'.join(sorted(emails)))
368
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar lowerCamelCase__ = TypeVar('T') class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Union[str, Any] , lowerCamelCase__ : T ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Dict = data _UpperCAmelCase : Node[T] | None = None def __str__( self : Any ) ->str: '''simple docstring''' return F"""{self.data}""" class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Tuple ) ->None: '''simple docstring''' _UpperCAmelCase : Node[T] | None = None def __iter__( self : List[str] ) ->Iterator[T]: '''simple docstring''' _UpperCAmelCase : Any = self.top while node: yield node.data _UpperCAmelCase : Dict = node.next def __str__( self : Dict ) ->str: '''simple docstring''' return "->".join([str(lowerCamelCase__ ) for item in self] ) def __len__( self : Optional[int] ) ->int: '''simple docstring''' return len(tuple(iter(self ) ) ) def lowerCAmelCase__ ( self : List[Any] ) ->bool: '''simple docstring''' return self.top is None def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : T ) ->None: '''simple docstring''' _UpperCAmelCase : List[Any] = Node(lowerCamelCase__ ) if not self.is_empty(): _UpperCAmelCase : Tuple = self.top _UpperCAmelCase : List[str] = node def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("pop from empty stack" ) assert isinstance(self.top , lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.top _UpperCAmelCase : Optional[Any] = self.top.next return pop_node.data def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("peek from empty stack" ) assert self.top is not None return self.top.data def lowerCAmelCase__ ( self : List[Any] ) ->None: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = None if __name__ == "__main__": from doctest import testmod testmod()
322
0
'''simple docstring''' import os from datetime import datetime as dt from github import Github lowerCamelCase__ = [ 'good first issue', 'feature request', 'wip', ] def __lowerCAmelCase (): _UpperCAmelCase : Optional[Any] = Github(os.environ["GITHUB_TOKEN"] ) _UpperCAmelCase : int = g.get_repo("huggingface/accelerate" ) _UpperCAmelCase : Tuple = repo.get_issues(state="open" ) for issue in open_issues: _UpperCAmelCase : Any = sorted([comment for comment in issue.get_comments()] , key=lambda __lowerCAmelCase : i.created_at , reverse=__lowerCAmelCase ) _UpperCAmelCase : List[Any] = comments[0] if len(__lowerCAmelCase ) > 0 else None _UpperCAmelCase : Union[str, Any] = dt.utcnow() _UpperCAmelCase : Union[str, Any] = (current_time - issue.updated_at).days _UpperCAmelCase : int = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state="closed" ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) if __name__ == "__main__": main()
369
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : int = "speech_to_text_2" lowerCAmelCase : str = ["past_key_values"] lowerCAmelCase : int = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[Any] , lowerCamelCase__ : Tuple=1_00_00 , lowerCamelCase__ : Any=6 , lowerCamelCase__ : Tuple=20_48 , lowerCamelCase__ : List[Any]=4 , lowerCamelCase__ : Tuple=0.0 , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Tuple="relu" , lowerCamelCase__ : Dict=2_56 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : List[Any]=0.0 , lowerCamelCase__ : Optional[int]=0.0 , lowerCamelCase__ : List[Any]=0.0_2 , lowerCamelCase__ : Tuple=2 , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Any=1 , lowerCamelCase__ : int=0 , lowerCamelCase__ : str=2 , lowerCamelCase__ : List[Any]=10_24 , **lowerCamelCase__ : str , ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = vocab_size _UpperCAmelCase : Optional[int] = d_model _UpperCAmelCase : List[Any] = decoder_ffn_dim _UpperCAmelCase : Any = decoder_layers _UpperCAmelCase : int = decoder_attention_heads _UpperCAmelCase : Any = dropout _UpperCAmelCase : List[Any] = attention_dropout _UpperCAmelCase : Optional[int] = activation_dropout _UpperCAmelCase : List[Any] = activation_function _UpperCAmelCase : int = init_std _UpperCAmelCase : Dict = decoder_layerdrop _UpperCAmelCase : str = use_cache _UpperCAmelCase : Union[str, Any] = decoder_layers _UpperCAmelCase : Optional[Any] = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCAmelCase : Any = max_target_positions super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return x if y == 0 else greatest_common_divisor(__lowerCAmelCase , x % y ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return (x * y) // greatest_common_divisor(__lowerCAmelCase , __lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase = 20 ): _UpperCAmelCase : List[Any] = 1 for i in range(1 , n + 1 ): _UpperCAmelCase : Dict = lcm(__lowerCAmelCase , __lowerCAmelCase ) return g if __name__ == "__main__": print(F'''{solution() = }''')
370
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = 'cuda' if torch.cuda.is_available() else 'cpu' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=100 , __lowerCAmelCase=" " ): _UpperCAmelCase : Any = text.split(__lowerCAmelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(__lowerCAmelCase ) , __lowerCAmelCase )] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase , _UpperCAmelCase : Dict = [], [] for title, text in zip(documents["title"] , documents["text"] ): if text is not None: for passage in split_text(__lowerCAmelCase ): titles.append(title if title is not None else "" ) texts.append(__lowerCAmelCase ) return {"title": titles, "text": texts} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : str = ctx_tokenizer( documents["title"] , documents["text"] , truncation=__lowerCAmelCase , padding="longest" , return_tensors="pt" )["input_ids"] _UpperCAmelCase : str = ctx_encoder(input_ids.to(device=__lowerCAmelCase ) , return_dict=__lowerCAmelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): ###################################### logger.info("Step 1 - Create the dataset" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way _UpperCAmelCase : Optional[int] = load_dataset( "csv" , data_files=[rag_example_args.csv_path] , split="train" , delimiter="\t" , column_names=["title", "text"] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words _UpperCAmelCase : Optional[int] = dataset.map(__lowerCAmelCase , batched=__lowerCAmelCase , num_proc=processing_args.num_proc ) # And compute the embeddings _UpperCAmelCase : Union[str, Any] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) _UpperCAmelCase : Dict = Features( {"text": Value("string" ), "title": Value("string" ), "embeddings": Sequence(Value("float32" ) )} ) # optional, save as float32 instead of float64 to save space _UpperCAmelCase : int = dataset.map( partial(__lowerCAmelCase , ctx_encoder=__lowerCAmelCase , ctx_tokenizer=__lowerCAmelCase ) , batched=__lowerCAmelCase , batch_size=processing_args.batch_size , features=__lowerCAmelCase , ) # And finally save your dataset _UpperCAmelCase : List[Any] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset" ) dataset.save_to_disk(__lowerCAmelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("Step 2 - Index the dataset" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search _UpperCAmelCase : Any = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("embeddings" , custom_index=__lowerCAmelCase ) # And save the index _UpperCAmelCase : List[str] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset_hnsw_index.faiss" ) dataset.get_index("embeddings" ).save(__lowerCAmelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" / "my_knowledge_dataset.csv" ) , metadata={"help": "Path to a tab-separated csv file with columns 'title' and 'text'"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Question that is passed as input to RAG. Default is 'What does Moses' rod turn into ?'."} , ) lowerCAmelCase : str = field( default="facebook/rag-sequence-nq" , metadata={"help": "The RAG model to use. Either 'facebook/rag-sequence-nq' or 'facebook/rag-token-nq'"} , ) lowerCAmelCase : str = field( default="facebook/dpr-ctx_encoder-multiset-base" , metadata={ "help": ( "The DPR context encoder model to use. Either 'facebook/dpr-ctx_encoder-single-nq-base' or" " 'facebook/dpr-ctx_encoder-multiset-base'" ) } , ) lowerCAmelCase : Optional[str] = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" ) , metadata={"help": "Path to a directory where the dataset passages and the index will be saved"} , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": "The number of processes to use to split the documents into passages. Default is single process." } , ) lowerCAmelCase : int = field( default=16 , metadata={ "help": "The batch size to use when computing the passages embeddings using the DPR context encoder." } , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : int = field( default=768 , metadata={"help": "The dimension of the embeddings to pass to the HNSW Faiss index."} , ) lowerCAmelCase : int = field( default=128 , metadata={ "help": ( "The number of bi-directional links created for every new element during the HNSW index construction." ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
322
0
import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments lowerCamelCase__ = logging.getLogger(__name__) @dataclass class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[float] = field( default=0.0 , metadata={"help": "The label smoothing epsilon to apply (if not zero)."} ) lowerCAmelCase : bool = field(default=UpperCAmelCase__ , metadata={"help": "Whether to SortishSamler or not."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Whether to use generate to calculate generative metrics (ROUGE, BLEU)."} ) lowerCAmelCase : bool = field(default=UpperCAmelCase__ , metadata={"help": "whether to use adafactor"} ) lowerCAmelCase : Optional[float] = field( default=UpperCAmelCase__ , metadata={"help": "Encoder layer dropout probability. Goes into model.config."} ) lowerCAmelCase : Optional[float] = field( default=UpperCAmelCase__ , metadata={"help": "Decoder layer dropout probability. Goes into model.config."} ) lowerCAmelCase : Optional[float] = field(default=UpperCAmelCase__ , metadata={"help": "Dropout probability. Goes into model.config."} ) lowerCAmelCase : Optional[float] = field( default=UpperCAmelCase__ , metadata={"help": "Attention dropout probability. Goes into model.config."} ) lowerCAmelCase : Optional[str] = field( default="linear" , metadata={"help": F"Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}"} , )
371
'''simple docstring''' import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCamelCase__ = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class lowerCAmelCase__ ( unittest.TestCase ): @classmethod def lowerCAmelCase__ ( cls : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : Tuple = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def lowerCAmelCase__ ( cls : Union[str, Any] ) ->int: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def lowerCAmelCase__ ( self : int ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ , repo_id="test-config" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : Dict = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ , repo_id="valid_org/test-config-org" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : int = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' CustomConfig.register_for_auto_class() _UpperCAmelCase : int = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) _UpperCAmelCase : str = AutoConfig.from_pretrained(F"""{USER}/test-dynamic-config""" , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : List[str] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated _UpperCAmelCase : Any = c.n_embd + 1 # int _UpperCAmelCase : List[Any] = c.resid_pdrop + 1.0 # float _UpperCAmelCase : Tuple = not c.scale_attn_weights # bool _UpperCAmelCase : List[Any] = c.summary_type + "foo" # str c.update_from_string( F"""n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}""" ) self.assertEqual(lowerCamelCase__ , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(lowerCamelCase__ , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(lowerCamelCase__ , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(lowerCamelCase__ , c.summary_type , "mismatch for key: summary_type" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = PretrainedConfig() _UpperCAmelCase : Tuple = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase__ , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) _UpperCAmelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase__ , lowerCamelCase__ )] if len(lowerCamelCase__ ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" F""" {', '.join(lowerCamelCase__ )}.""" ) def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): # config is in subfolder, the following should not work without specifying the subfolder _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = mock.Mock() _UpperCAmelCase : List[str] = 5_00 _UpperCAmelCase : Dict = {} _UpperCAmelCase : Tuple = HTTPError _UpperCAmelCase : Any = {} # Download this model to make sure it's in the cache. _UpperCAmelCase : int = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=lowerCamelCase__ ) as mock_head: _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = AutoConfig.from_pretrained("bert-base-cased" ) _UpperCAmelCase : str = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 _UpperCAmelCase : Optional[int] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 _UpperCAmelCase : Dict = ["config.42.0.0.json"] _UpperCAmelCase : Union[str, Any] = 7_68 configuration.save_pretrained(lowerCamelCase__ ) shutil.move(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , os.path.join(lowerCamelCase__ , "config.42.0.0.json" ) ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = "hf-internal-testing/test-two-configs" import transformers as new_transformers _UpperCAmelCase : Any = "v4.0.0" _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers _UpperCAmelCase : List[Any] = "v3.0.0" _UpperCAmelCase : int = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(old_configuration.hidden_size , 7_68 )
322
0
'''simple docstring''' import collections import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = '▁' lowerCamelCase__ = {'vocab_file': 'prophetnet.tokenizer'} lowerCamelCase__ = { 'vocab_file': { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/prophetnet.tokenizer' ), } } lowerCamelCase__ = { 'microsoft/xprophetnet-large-wiki100-cased': {'do_lower_case': False}, } lowerCamelCase__ = { 'microsoft/xprophetnet-large-wiki100-cased': 512, } def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Any = collections.OrderedDict() with open(__lowerCAmelCase , "r" , encoding="utf-8" ) as reader: _UpperCAmelCase : Any = reader.readlines() for index, token in enumerate(__lowerCAmelCase ): _UpperCAmelCase : str = token.rstrip("\n" ) _UpperCAmelCase : Any = index return vocab class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = VOCAB_FILES_NAMES lowerCAmelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase : Tuple = ["input_ids", "attention_mask"] def __init__( self : Tuple , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[int]="[SEP]" , lowerCamelCase__ : Dict="[SEP]" , lowerCamelCase__ : List[str]="[SEP]" , lowerCamelCase__ : List[Any]="[UNK]" , lowerCamelCase__ : Optional[int]="[PAD]" , lowerCamelCase__ : List[str]="[CLS]" , lowerCamelCase__ : int="[MASK]" , lowerCamelCase__ : Optional[Dict[str, Any]] = None , **lowerCamelCase__ : int , ) ->None: '''simple docstring''' _UpperCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase__ , ) try: import sentencepiece as spm except ImportError: logger.warning( "You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece" " pip install sentencepiece" ) raise _UpperCAmelCase : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # put special tokens and [unused] tokens into the vocab _UpperCAmelCase : List[str] = {"[PAD]": 0, "[CLS]": 1, "[SEP]": 2, "[UNK]": 3, "[MASK]": 4} for i in range(10 ): _UpperCAmelCase : List[Any] = F"""[unused{i}]""" _UpperCAmelCase : Union[str, Any] = 5 + i # The first "real" token "," has position 15 in the embedding vocab and position 3 in the spm vocab _UpperCAmelCase : Tuple = 12 _UpperCAmelCase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} for k in self.fairseq_tokens_to_ids.keys(): self.unique_no_split_tokens.append(lowerCamelCase__ ) def __getstate__( self : Dict ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Dict = self.__dict__.copy() _UpperCAmelCase : int = None return state def __setstate__( self : List[Any] , lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Any = d try: import sentencepiece as spm except ImportError: logger.warning( "You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece" " pip install sentencepiece" ) raise # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): _UpperCAmelCase : Any = {} _UpperCAmelCase : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : List[int] , lowerCamelCase__ : Optional[List[int]] = None , lowerCamelCase__ : bool = False ) ->List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return ([0] * len(lowerCamelCase__ )) + [1] return ([0] * len(lowerCamelCase__ )) + [1] + ([0] * len(lowerCamelCase__ )) + [1] def lowerCAmelCase__ ( self : str , lowerCamelCase__ : List[int] , lowerCamelCase__ : Optional[List[int]] = None ) ->List[int]: '''simple docstring''' _UpperCAmelCase : Tuple = [self.sep_token_id] if token_ids_a is None: return len(token_ids_a + sep ) * [0] return len(token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def lowerCAmelCase__ ( self : Any ) ->Optional[int]: '''simple docstring''' return len(self.sp_model ) + self.fairseq_offset def lowerCAmelCase__ ( self : Any ) ->Dict: '''simple docstring''' _UpperCAmelCase : Optional[int] = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCAmelCase__ ( self : int , lowerCamelCase__ : str ) ->str: '''simple docstring''' return self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : List[str] ) ->List[str]: '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _UpperCAmelCase : str = self.sp_model.PieceToId(lowerCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : int ) ->List[Any]: '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : Union[str, Any] ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Tuple = "".join(lowerCamelCase__ ).replace(lowerCamelCase__ , " " ).strip() return out_string def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : str , lowerCamelCase__ : Optional[str] = None ) ->Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _UpperCAmelCase : Union[str, Any] = os.path.join( lowerCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , "wb" ) as fi: _UpperCAmelCase : int = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,) def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : List[int] , lowerCamelCase__ : Optional[List[int]] = None ) ->List[int]: '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.sep_token_id] _UpperCAmelCase : Any = [self.sep_token_id] return token_ids_a + sep + token_ids_a + sep
350
'''simple docstring''' from manim import * class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : List[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Optional[Any] = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Dict = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("CPU" , font_size=24 ) _UpperCAmelCase : Any = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(1 )] _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("GPU" , font_size=24 ) _UpperCAmelCase : str = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) gpu.align_to(lowerCamelCase__ , lowerCamelCase__ ) gpu.set_x(gpu.get_x() - 1 ) self.add(lowerCamelCase__ ) _UpperCAmelCase : List[str] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Any = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Model" , font_size=24 ) _UpperCAmelCase : Tuple = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.play( Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , ) _UpperCAmelCase : int = MarkupText( F"""First, an empty model skeleton is loaded\ninto <span fgcolor='{YELLOW}'>memory</span> without using much RAM.""" , font_size=24 , ) _UpperCAmelCase : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ , run_time=2.5 ) , Write(lowerCamelCase__ ) , Write(lowerCamelCase__ ) ) self.add(lowerCamelCase__ ) _UpperCAmelCase : int = [] _UpperCAmelCase : List[str] = [] _UpperCAmelCase : Dict = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : int = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ , opacity=0.7 ) cpu_target.move_to(lowerCamelCase__ ) cpu_target.generate_target() _UpperCAmelCase : Dict = 0.4_6 / 4 _UpperCAmelCase : Any = 0.4_6 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowerCamelCase__ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=lowerCamelCase__ , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=lowerCamelCase__ , buff=0.0 ) cpu_targs.append(lowerCamelCase__ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(lowerCamelCase__ ) ) second_animations.append(MoveToTarget(lowerCamelCase__ , run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(*lowerCamelCase__ ) self.wait()
322
0
'''simple docstring''' from typing import Any class lowerCAmelCase__ : def __init__( self : List[Any] , lowerCamelCase__ : Any ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Optional[Any] = data _UpperCAmelCase : Tuple = None class lowerCAmelCase__ : def __init__( self : str ) ->Dict: '''simple docstring''' _UpperCAmelCase : Optional[int] = None def lowerCAmelCase__ ( self : Any ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = self.head while temp is not None: print(temp.data , end=" " ) _UpperCAmelCase : int = temp.next print() def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Any ) ->Any: '''simple docstring''' _UpperCAmelCase : Tuple = Node(lowerCamelCase__ ) _UpperCAmelCase : Any = self.head _UpperCAmelCase : Union[str, Any] = new_node def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' if node_data_a == node_data_a: return else: _UpperCAmelCase : Union[str, Any] = self.head while node_a is not None and node_a.data != node_data_a: _UpperCAmelCase : Union[str, Any] = node_a.next _UpperCAmelCase : List[str] = self.head while node_a is not None and node_a.data != node_data_a: _UpperCAmelCase : Optional[Any] = node_a.next if node_a is None or node_a is None: return _UpperCAmelCase : Any = node_a.data, node_a.data if __name__ == "__main__": lowerCamelCase__ = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('After swapping') ll.print_list()
351
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1_024 , __lowerCAmelCase=1_024 , __lowerCAmelCase=False , **__lowerCAmelCase ): _UpperCAmelCase : Any = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : List[str] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="train" , **__lowerCAmelCase ) _UpperCAmelCase : Dict = tok.pad_token_id def get_lens(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = tqdm( DataLoader(__lowerCAmelCase , batch_size=512 , num_workers=8 , shuffle=__lowerCAmelCase , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) _UpperCAmelCase : List[str] = [] for batch in dl: _UpperCAmelCase : Any = batch["input_ids"].ne(__lowerCAmelCase ).sum(1 ).tolist() _UpperCAmelCase : Tuple = batch["labels"].ne(__lowerCAmelCase ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__lowerCAmelCase , __lowerCAmelCase ): max_lens.append(max(__lowerCAmelCase , __lowerCAmelCase ) ) else: max_lens.extend(__lowerCAmelCase ) return max_lens _UpperCAmelCase : Dict = get_lens(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="val" , **__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = get_lens(__lowerCAmelCase ) pickle_save(__lowerCAmelCase , train_ds.len_file ) pickle_save(__lowerCAmelCase , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
322
0
'''simple docstring''' import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : int ) ->str: '''simple docstring''' _UpperCAmelCase : Optional[Any] = tempfile.mkdtemp() _UpperCAmelCase : List[str] = 5 # Realm tok _UpperCAmelCase : Optional[Any] = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "test", "question", "this", "is", "the", "first", "second", "third", "fourth", "fifth", "record", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] _UpperCAmelCase : List[Any] = os.path.join(self.tmpdirname , "realm_tokenizer" ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) _UpperCAmelCase : Any = os.path.join(lowerCamelCase__ , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) _UpperCAmelCase : Optional[int] = os.path.join(self.tmpdirname , "realm_block_records" ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Any ) ->RealmTokenizer: '''simple docstring''' return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , "realm_tokenizer" ) ) def lowerCAmelCase__ ( self : int ) ->str: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def lowerCAmelCase__ ( self : List[str] ) ->int: '''simple docstring''' _UpperCAmelCase : Optional[Any] = RealmConfig(num_block_records=self.num_block_records ) return config def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = Dataset.from_dict( { "id": ["0", "1"], "question": ["foo", "bar"], "answers": [["Foo", "Bar"], ["Bar"]], } ) return dataset def lowerCAmelCase__ ( self : str ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = np.array( [ b"This is the first record", b"This is the second record", b"This is the third record", b"This is the fourth record", b"This is the fifth record", b"This is a longer longer longer record", ] , dtype=lowerCamelCase__ , ) return block_records def lowerCAmelCase__ ( self : Dict ) ->int: '''simple docstring''' _UpperCAmelCase : Tuple = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def lowerCAmelCase__ ( self : Tuple ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = self.get_config() _UpperCAmelCase : Tuple = self.get_dummy_retriever() _UpperCAmelCase : Any = retriever.tokenizer _UpperCAmelCase : Union[str, Any] = np.array([0, 3] , dtype="long" ) _UpperCAmelCase : Tuple = tokenizer(["Test question"] ).input_ids _UpperCAmelCase : Any = tokenizer( ["the fourth"] , add_special_tokens=lowerCamelCase__ , return_token_type_ids=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , ).input_ids _UpperCAmelCase : Any = config.reader_seq_len _UpperCAmelCase : Optional[int] = retriever( lowerCamelCase__ , lowerCamelCase__ , answer_ids=lowerCamelCase__ , max_length=lowerCamelCase__ , return_tensors="np" ) self.assertEqual(len(lowerCamelCase__ ) , 2 ) self.assertEqual(len(lowerCamelCase__ ) , 2 ) self.assertEqual(len(lowerCamelCase__ ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 10) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 10) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 10) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ["[CLS]", "test", "question", "[SEP]", "this", "is", "the", "first", "record", "[SEP]"] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ["[CLS]", "test", "question", "[SEP]", "this", "is", "the", "fourth", "record", "[SEP]"] , ) def lowerCAmelCase__ ( self : int ) ->int: '''simple docstring''' _UpperCAmelCase : Any = self.get_config() _UpperCAmelCase : int = self.get_dummy_retriever() _UpperCAmelCase : Optional[int] = retriever.tokenizer _UpperCAmelCase : Any = np.array([0, 3, 5] , dtype="long" ) _UpperCAmelCase : List[str] = tokenizer(["Test question"] ).input_ids _UpperCAmelCase : Dict = tokenizer( ["the fourth", "longer longer"] , add_special_tokens=lowerCamelCase__ , return_token_type_ids=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , ).input_ids _UpperCAmelCase : Optional[Any] = config.reader_seq_len _UpperCAmelCase : int = retriever( lowerCamelCase__ , lowerCamelCase__ , answer_ids=lowerCamelCase__ , max_length=lowerCamelCase__ , return_tensors="np" ) self.assertEqual([False, True, True] , lowerCamelCase__ ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , lowerCamelCase__ ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Union[str, Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , "realm_block_records" ) ) # Test local path _UpperCAmelCase : Optional[int] = retriever.from_pretrained(os.path.join(self.tmpdirname , "realm_block_records" ) ) self.assertEqual(retriever.block_records[0] , b"This is the first record" ) # Test mocked remote path with patch("transformers.models.realm.retrieval_realm.hf_hub_download" ) as mock_hf_hub_download: _UpperCAmelCase : List[Any] = os.path.join( os.path.join(self.tmpdirname , "realm_block_records" ) , _REALM_BLOCK_RECORDS_FILENAME ) _UpperCAmelCase : int = RealmRetriever.from_pretrained("google/realm-cc-news-pretrained-openqa" ) self.assertEqual(retriever.block_records[0] , b"This is the first record" )
352
'''simple docstring''' import pytest lowerCamelCase__ = '__dummy_dataset1__' lowerCamelCase__ = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = dataset_loading_script_name _UpperCAmelCase : Any = tmp_path / "datasets" / script_name script_dir.mkdir(parents=__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = script_dir / F"""{script_name}.py""" with open(__lowerCAmelCase , "w" ) as f: f.write(__lowerCAmelCase ) return str(__lowerCAmelCase )
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ) -> Union[str, Any]: _UpperCAmelCase : Tuple = [0] * len(__lowerCAmelCase ) _UpperCAmelCase : int = [] _UpperCAmelCase : Optional[Any] = [] _UpperCAmelCase : Optional[Any] = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__lowerCAmelCase ) ): if indegree[i] == 0: queue.append(__lowerCAmelCase ) while queue: _UpperCAmelCase : Optional[int] = queue.pop(0 ) cnt += 1 topo.append(__lowerCAmelCase ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(__lowerCAmelCase ) if cnt != len(__lowerCAmelCase ): print("Cycle exists" ) else: print(__lowerCAmelCase ) # Adjacency List of Graph lowerCamelCase__ = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
353
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase__ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCamelCase__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCamelCase__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCamelCase__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"] , reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ] , ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] ) ->int: '''simple docstring''' import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : int=0.9 , lowerCamelCase__ : Dict=3 , lowerCamelCase__ : Dict=0.5 ) ->Any: '''simple docstring''' if NLTK_VERSION >= version.Version("3.6.5" ): _UpperCAmelCase : Dict = [ meteor_score.single_meteor_score( word_tokenize(lowerCamelCase__ ) , word_tokenize(lowerCamelCase__ ) , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] else: _UpperCAmelCase : Optional[int] = [ meteor_score.single_meteor_score(lowerCamelCase__ , lowerCamelCase__ , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] return {"meteor": np.mean(lowerCamelCase__ )}
322
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { 'configuration_conditional_detr': [ 'CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ConditionalDetrConfig', 'ConditionalDetrOnnxConfig', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ['ConditionalDetrFeatureExtractor'] lowerCamelCase__ = ['ConditionalDetrImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ 'CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST', 'ConditionalDetrForObjectDetection', 'ConditionalDetrForSegmentation', 'ConditionalDetrModel', 'ConditionalDetrPreTrainedModel', ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
354
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline lowerCamelCase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : int ) ->str: '''simple docstring''' if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Union[str, Any] = [label.strip() for label in labels.split("," ) if label.strip()] return labels def __call__( self : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] ) ->str: '''simple docstring''' if len(lowerCamelCase__ ) == 0 or len(lowerCamelCase__ ) == 0: raise ValueError("You must include at least one label and at least one sequence." ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( "The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. " "Make sure the passed template includes formatting syntax such as {{}} where the label should go." ).format(lowerCamelCase__ ) ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Optional[Any] = [sequences] _UpperCAmelCase : int = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(lowerCamelCase__ )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(UpperCAmelCase__ ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Union[str, Any] , lowerCamelCase__ : Optional[Any]=ZeroShotClassificationArgumentHandler() , *lowerCamelCase__ : List[str] , **lowerCamelCase__ : Any ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = args_parser super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) if self.entailment_id == -1: logger.warning( "Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to " "-1. Define a descriptive label2id mapping in the model config to ensure correct outputs." ) @property def lowerCAmelCase__ ( self : Any ) ->Union[str, Any]: '''simple docstring''' for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("entail" ): return ind return -1 def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Tuple , lowerCamelCase__ : int=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : str=TruncationStrategy.ONLY_FIRST , **lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : int = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( "Tokenizer was not supporting padding necessary for zero-shot, attempting to use " " `pad_token=eos_token`" ) _UpperCAmelCase : Optional[Any] = self.tokenizer.eos_token try: _UpperCAmelCase : List[str] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , ) except Exception as e: if "too short" in str(lowerCamelCase__ ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. _UpperCAmelCase : List[Any] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def lowerCAmelCase__ ( self : int , **lowerCamelCase__ : Union[str, Any] ) ->Tuple: '''simple docstring''' if kwargs.get("multi_class" , lowerCamelCase__ ) is not None: _UpperCAmelCase : int = kwargs["multi_class"] logger.warning( "The `multi_class` argument has been deprecated and renamed to `multi_label`. " "`multi_class` will be removed in a future version of Transformers." ) _UpperCAmelCase : Dict = {} if "candidate_labels" in kwargs: _UpperCAmelCase : List[Any] = self._args_parser._parse_labels(kwargs["candidate_labels"] ) if "hypothesis_template" in kwargs: _UpperCAmelCase : Dict = kwargs["hypothesis_template"] _UpperCAmelCase : List[str] = {} if "multi_label" in kwargs: _UpperCAmelCase : Optional[Any] = kwargs["multi_label"] return preprocess_params, {}, postprocess_params def __call__( self : int , lowerCamelCase__ : Union[str, List[str]] , *lowerCamelCase__ : str , **lowerCamelCase__ : Optional[Any] , ) ->Optional[int]: '''simple docstring''' if len(lowerCamelCase__ ) == 0: pass elif len(lowerCamelCase__ ) == 1 and "candidate_labels" not in kwargs: _UpperCAmelCase : int = args[0] else: raise ValueError(F"""Unable to understand extra arguments {args}""" ) return super().__call__(lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Any=None , lowerCamelCase__ : str="This example is {}." ) ->Tuple: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Optional[int] = self._args_parser(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for i, (candidate_label, sequence_pair) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): _UpperCAmelCase : Optional[int] = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(lowerCamelCase__ ) - 1, **model_input, } def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Optional[int] ) ->int: '''simple docstring''' _UpperCAmelCase : Dict = inputs["candidate_label"] _UpperCAmelCase : Optional[int] = inputs["sequence"] _UpperCAmelCase : Dict = {k: inputs[k] for k in self.tokenizer.model_input_names} _UpperCAmelCase : List[Any] = self.model(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = { "candidate_label": candidate_label, "sequence": sequence, "is_last": inputs["is_last"], **outputs, } return model_outputs def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple=False ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = [outputs["candidate_label"] for outputs in model_outputs] _UpperCAmelCase : Any = [outputs["sequence"] for outputs in model_outputs] _UpperCAmelCase : Optional[int] = np.concatenate([output["logits"].numpy() for output in model_outputs] ) _UpperCAmelCase : Optional[Any] = logits.shape[0] _UpperCAmelCase : Any = len(lowerCamelCase__ ) _UpperCAmelCase : str = N // n _UpperCAmelCase : str = logits.reshape((num_sequences, n, -1) ) if multi_label or len(lowerCamelCase__ ) == 1: # softmax over the entailment vs. contradiction dim for each label independently _UpperCAmelCase : int = self.entailment_id _UpperCAmelCase : List[Any] = -1 if entailment_id == 0 else 0 _UpperCAmelCase : str = reshaped_outputs[..., [contradiction_id, entailment_id]] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : str = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels _UpperCAmelCase : int = reshaped_outputs[..., self.entailment_id] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
322
0
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1_024 , __lowerCAmelCase=1_024 , __lowerCAmelCase=False , **__lowerCAmelCase ): _UpperCAmelCase : Any = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : List[str] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="train" , **__lowerCAmelCase ) _UpperCAmelCase : Dict = tok.pad_token_id def get_lens(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = tqdm( DataLoader(__lowerCAmelCase , batch_size=512 , num_workers=8 , shuffle=__lowerCAmelCase , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) _UpperCAmelCase : List[str] = [] for batch in dl: _UpperCAmelCase : Any = batch["input_ids"].ne(__lowerCAmelCase ).sum(1 ).tolist() _UpperCAmelCase : Tuple = batch["labels"].ne(__lowerCAmelCase ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__lowerCAmelCase , __lowerCAmelCase ): max_lens.append(max(__lowerCAmelCase , __lowerCAmelCase ) ) else: max_lens.extend(__lowerCAmelCase ) return max_lens _UpperCAmelCase : Dict = get_lens(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="val" , **__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = get_lens(__lowerCAmelCase ) pickle_save(__lowerCAmelCase , train_ds.len_file ) pickle_save(__lowerCAmelCase , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
355
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase = 4_000_000 ): _UpperCAmelCase : List[Any] = [] _UpperCAmelCase , _UpperCAmelCase : Dict = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__lowerCAmelCase ) _UpperCAmelCase , _UpperCAmelCase : Any = b, a + b return sum(__lowerCAmelCase ) if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): return credit_card_number.startswith(("34", "35", "37", "4", "5", "6") ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[str] = credit_card_number _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : str = len(__lowerCAmelCase ) - 2 for i in range(__lowerCAmelCase , -1 , -2 ): # double the value of every second digit _UpperCAmelCase : Tuple = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 _UpperCAmelCase : Tuple = cc_number[:i] + str(__lowerCAmelCase ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(__lowerCAmelCase ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Optional[int] = F"""{credit_card_number} is an invalid credit card number because""" if not credit_card_number.isdigit(): print(F"""{error_message} it has nonnumerical characters.""" ) return False if not 13 <= len(__lowerCAmelCase ) <= 16: print(F"""{error_message} of its length.""" ) return False if not validate_initial_digits(__lowerCAmelCase ): print(F"""{error_message} of its first two digits.""" ) return False if not luhn_validation(__lowerCAmelCase ): print(F"""{error_message} it fails the Luhn check.""" ) return False print(F"""{credit_card_number} is a valid credit card number.""" ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number('4111111111111111') validate_credit_card_number('32323')
356
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[str]=13 , lowerCamelCase__ : Optional[Any]=7 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : int=99 , lowerCamelCase__ : int=32 , lowerCamelCase__ : List[str]=5 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : Optional[int]=37 , lowerCamelCase__ : Tuple="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Union[str, Any]=0.1 , lowerCamelCase__ : Optional[int]=5_12 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : str=2 , lowerCamelCase__ : Union[str, Any]=0.0_2 , lowerCamelCase__ : Tuple=4 , ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = parent _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Optional[int] = seq_length _UpperCAmelCase : int = is_training _UpperCAmelCase : Dict = use_attention_mask _UpperCAmelCase : Optional[Any] = use_token_type_ids _UpperCAmelCase : int = use_labels _UpperCAmelCase : Optional[int] = vocab_size _UpperCAmelCase : Any = hidden_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : List[Any] = num_attention_heads _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : int = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob _UpperCAmelCase : Union[str, Any] = max_position_embeddings _UpperCAmelCase : Tuple = type_vocab_size _UpperCAmelCase : List[Any] = type_sequence_label_size _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Dict = num_choices def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Dict = None if self.use_attention_mask: _UpperCAmelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCAmelCase : Union[str, Any] = None if self.use_token_type_ids: _UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCAmelCase : int = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = config_and_inputs _UpperCAmelCase : str = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Optional[int] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase__ ( self : Optional[int] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = FlaxAlbertModelTester(self ) @slow def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCAmelCase : List[str] = model_class_name.from_pretrained("albert-base-v2" ) _UpperCAmelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = FlaxAlbertModel.from_pretrained("albert-base-v2" ) _UpperCAmelCase : List[Any] = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) _UpperCAmelCase : Optional[int] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCAmelCase : Dict = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )[0] _UpperCAmelCase : List[Any] = (1, 11, 7_68) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : str = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase__ , atol=1E-4 ) )
322
0
'''simple docstring''' import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput lowerCamelCase__ : Optional[Any] = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , *lowerCamelCase__ : str , lowerCamelCase__ : Dict=None , lowerCamelCase__ : Optional[Any]=None , lowerCamelCase__ : str=None , **lowerCamelCase__ : Optional[int] ) ->Optional[int]: '''simple docstring''' super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = eval_examples _UpperCAmelCase : List[str] = post_process_function _UpperCAmelCase : List[Any] = quant_trainer_args _UpperCAmelCase : List[str] = 1_28 # default number of calibration samples def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Optional[Any]=None ) ->Any: '''simple docstring''' if calib_dataset is None and self.calib_dataset is None: raise ValueError("Trainer: calibration requires an calib_dataset." ) _UpperCAmelCase : Dict = calib_dataset if calib_dataset is not None else self.calib_dataset _UpperCAmelCase : Union[str, Any] = self._remove_unused_columns(lowerCamelCase__ , description="Calibration" ) return DataLoader( lowerCamelCase__ , batch_size=self.args.eval_batch_size , collate_fn=self.data_collator , drop_last=self.args.dataloader_drop_last , num_workers=self.args.dataloader_num_workers , pin_memory=self.args.dataloader_pin_memory , shuffle=lowerCamelCase__ , ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : List[str]=None ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.train_dataset if calib_dataset is None else calib_dataset _UpperCAmelCase : int = self.get_calib_dataloader(lowerCamelCase__ ) _UpperCAmelCase : str = self.model quant_trainer.configure_model(lowerCamelCase__ , self.quant_trainer_args , calib=lowerCamelCase__ ) model.eval() quant_trainer.enable_calibration(lowerCamelCase__ ) logger.info("***** Running calibration *****" ) logger.info(F""" Num examples = {self.calib_num}""" ) logger.info(F""" Batch size = {calib_dataloader.batch_size}""" ) for step, inputs in enumerate(lowerCamelCase__ ): # Prediction step _UpperCAmelCase : Any = self.prediction_step(lowerCamelCase__ , lowerCamelCase__ , prediction_loss_only=lowerCamelCase__ ) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(lowerCamelCase__ , self.quant_trainer_args ) _UpperCAmelCase : Dict = model def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : Tuple=None , lowerCamelCase__ : Any=None , lowerCamelCase__ : int=None , lowerCamelCase__ : str = "eval" ) ->Dict: '''simple docstring''' _UpperCAmelCase : Any = self.eval_dataset if eval_dataset is None else eval_dataset _UpperCAmelCase : Tuple = self.get_eval_dataloader(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. _UpperCAmelCase : List[str] = self.compute_metrics _UpperCAmelCase : int = None _UpperCAmelCase : Optional[int] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _UpperCAmelCase : Tuple = eval_loop( lowerCamelCase__ , description="Evaluation" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowerCamelCase__ , ) finally: _UpperCAmelCase : Tuple = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: _UpperCAmelCase : str = self.post_process_function(lowerCamelCase__ , lowerCamelCase__ , output.predictions ) _UpperCAmelCase : Optional[int] = self.compute_metrics(lowerCamelCase__ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): _UpperCAmelCase : List[Any] = metrics.pop(lowerCamelCase__ ) self.log(lowerCamelCase__ ) else: _UpperCAmelCase : List[str] = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) _UpperCAmelCase : str = self.callback_handler.on_evaluate(self.args , self.state , self.control , lowerCamelCase__ ) return metrics def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : List[Any]=None , lowerCamelCase__ : str = "test" ) ->Any: '''simple docstring''' _UpperCAmelCase : Any = self.get_test_dataloader(lowerCamelCase__ ) # Temporarily disable metric computation, we will do it in the loop here. _UpperCAmelCase : Dict = self.compute_metrics _UpperCAmelCase : Any = None _UpperCAmelCase : Tuple = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _UpperCAmelCase : List[Any] = eval_loop( lowerCamelCase__ , description="Prediction" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowerCamelCase__ , ) finally: _UpperCAmelCase : Union[str, Any] = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output _UpperCAmelCase : Optional[Any] = self.post_process_function(lowerCamelCase__ , lowerCamelCase__ , output.predictions , "predict" ) _UpperCAmelCase : Tuple = self.compute_metrics(lowerCamelCase__ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): _UpperCAmelCase : str = metrics.pop(lowerCamelCase__ ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Tuple="./" ) ->Any: '''simple docstring''' _UpperCAmelCase : Any = self.eval_dataset _UpperCAmelCase : Tuple = self.get_eval_dataloader(lowerCamelCase__ ) _UpperCAmelCase : Tuple = next(iter(lowerCamelCase__ ) ) # saving device - to make it consistent _UpperCAmelCase : Dict = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) # convert to tuple _UpperCAmelCase : List[Any] = tuple(v.to(lowerCamelCase__ ) for k, v in batch.items() ) logger.info("Converting model to be onnx compatible" ) from pytorch_quantization.nn import TensorQuantizer _UpperCAmelCase : Dict = True _UpperCAmelCase : str = self.model.to(lowerCamelCase__ ) model.eval() model.float() _UpperCAmelCase : int = model.module if hasattr(lowerCamelCase__ , "module" ) else model quant_trainer.configure_model(lowerCamelCase__ , self.quant_trainer_args ) _UpperCAmelCase : Optional[int] = os.path.join(lowerCamelCase__ , "model.onnx" ) logger.info(F"""exporting model to {output_model_file}""" ) _UpperCAmelCase : List[Any] = {0: "batch_size", 1: "seq_len"} torch.onnx.export( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , export_params=lowerCamelCase__ , opset_version=13 , do_constant_folding=lowerCamelCase__ , input_names=["input_ids", "attention_mask", "token_type_ids"] , output_names=["output_start_logits", "output_end_logits"] , dynamic_axes={ "input_ids": axes, "attention_mask": axes, "token_type_ids": axes, "output_start_logits": axes, "output_end_logits": axes, } , verbose=lowerCamelCase__ , ) logger.info("onnx export finished" )
357
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-classification/requirements.txt') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "rb" ) as f: _UpperCAmelCase : List[str] = Image.open(__lowerCAmelCase ) return im.convert("RGB" ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={ "help": "Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub)." } , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the training data."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the validation data."} ) lowerCAmelCase : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default="google/vit-base-patch16-224-in21k" , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCAmelCase__ )} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) lowerCAmelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCAmelCase : str = field(default=UpperCAmelCase__ , metadata={"help": "Name or path of preprocessor config."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = torch.stack([example["pixel_values"] for example in examples] ) _UpperCAmelCase : Tuple = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def __lowerCAmelCase (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , __lowerCAmelCase , __lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(__lowerCAmelCase ) transformers.utils.logging.set_verbosity(__lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _UpperCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCAmelCase : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCAmelCase : List[Any] = {} if data_args.train_dir is not None: _UpperCAmelCase : str = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCAmelCase : Optional[Any] = os.path.join(data_args.validation_dir , "**" ) _UpperCAmelCase : Any = load_dataset( "imagefolder" , data_files=__lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCAmelCase : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __lowerCAmelCase ) and data_args.train_val_split > 0.0: _UpperCAmelCase : List[Any] = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCAmelCase : List[str] = split["train"] _UpperCAmelCase : Union[str, Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase : Optional[int] = dataset["train"].features["labels"].names _UpperCAmelCase , _UpperCAmelCase : int = {}, {} for i, label in enumerate(__lowerCAmelCase ): _UpperCAmelCase : int = str(__lowerCAmelCase ) _UpperCAmelCase : str = label # Load the accuracy metric from the datasets package _UpperCAmelCase : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(__lowerCAmelCase ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowerCAmelCase ) , labelaid=__lowerCAmelCase , idalabel=__lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCAmelCase : Dict = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCAmelCase : int = image_processor.size["shortest_edge"] else: _UpperCAmelCase : int = (image_processor.size["height"], image_processor.size["width"]) _UpperCAmelCase : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCAmelCase : Optional[int] = Compose( [ RandomResizedCrop(__lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCAmelCase : Union[str, Any] = Compose( [ Resize(__lowerCAmelCase ), CenterCrop(__lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCAmelCase : Dict = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(__lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCAmelCase : Optional[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(__lowerCAmelCase ) # Initalize our trainer _UpperCAmelCase : Union[str, Any] = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=__lowerCAmelCase , tokenizer=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : Any = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : int = last_checkpoint _UpperCAmelCase : Dict = trainer.train(resume_from_checkpoint=__lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowerCAmelCase ) trainer.save_metrics("eval" , __lowerCAmelCase ) # Write model card and (optionally) push to hub _UpperCAmelCase : int = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowerCAmelCase ) else: trainer.create_model_card(**__lowerCAmelCase ) if __name__ == "__main__": main()
322
0
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowerCamelCase__ = { 'cola': 2, 'mnli': 3, 'mrpc': 2, 'sst-2': 2, 'sts-b': 1, 'qqp': 2, 'qnli': 2, 'rte': 2, 'wnli': 2, } logging.set_verbosity_info() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None ): # Initialise PyTorch model _UpperCAmelCase : List[str] = XLNetConfig.from_json_file(__lowerCAmelCase ) _UpperCAmelCase : List[str] = finetuning_task.lower() if finetuning_task is not None else "" if finetuning_task in GLUE_TASKS_NUM_LABELS: print(F"""Building PyTorch XLNetForSequenceClassification model from configuration: {config}""" ) _UpperCAmelCase : List[str] = finetuning_task _UpperCAmelCase : Tuple = GLUE_TASKS_NUM_LABELS[finetuning_task] _UpperCAmelCase : str = XLNetForSequenceClassification(__lowerCAmelCase ) elif "squad" in finetuning_task: _UpperCAmelCase : Tuple = finetuning_task _UpperCAmelCase : Optional[Any] = XLNetForQuestionAnswering(__lowerCAmelCase ) else: _UpperCAmelCase : Any = XLNetLMHeadModel(__lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # Save pytorch-model _UpperCAmelCase : Dict = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : List[Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) print(F"""Save PyTorch model to {os.path.abspath(__lowerCAmelCase )}""" ) torch.save(model.state_dict() , __lowerCAmelCase ) print(F"""Save configuration file to {os.path.abspath(__lowerCAmelCase )}""" ) with open(__lowerCAmelCase , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--xlnet_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained XLNet model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--finetuning_task', default=None, type=str, help='Name of a task on which the XLNet TensorFlow model was fine-tuned', ) lowerCamelCase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
358
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase__ = logging.get_logger(__name__) # General docstring lowerCamelCase__ = 'RegNetConfig' # Base docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = [1, 1_088, 7, 7] # Image classification docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = 'tabby, tabby cat' lowerCamelCase__ = [ 'facebook/regnet-y-040', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 3 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[str] = "relu" , **lowerCamelCase__ : Tuple , ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _UpperCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _UpperCAmelCase : Dict = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=lowerCamelCase__ , strides=lowerCamelCase__ , padding="VALID" , groups=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" , ) _UpperCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) _UpperCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.convolution(self.padding(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = self.normalization(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = config.num_channels _UpperCAmelCase : Any = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = shape_list(lowerCamelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _UpperCAmelCase : Optional[Any] = tf.transpose(lowerCamelCase__ , perm=(0, 2, 3, 1) ) _UpperCAmelCase : List[Any] = self.embedder(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : int = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=1 , strides=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" ) _UpperCAmelCase : Any = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False ) ->tf.Tensor: '''simple docstring''' return self.normalization(self.convolution(lowerCamelCase__ ) , training=lowerCamelCase__ ) class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : int , **lowerCamelCase__ : Optional[int] ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) _UpperCAmelCase : int = [ tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.pooler(lowerCamelCase__ ) for layer_module in self.attention: _UpperCAmelCase : str = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : Any ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = in_channels != out_channels or stride != 1 _UpperCAmelCase : List[str] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : List[str] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _UpperCAmelCase : Optional[int] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.2" ), ] _UpperCAmelCase : Union[str, Any] = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = hidden_state for layer_module in self.layers: _UpperCAmelCase : List[Any] = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : List[Any] , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Optional[int] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : Union[str, Any] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) _UpperCAmelCase : List[Any] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(lowerCamelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.3" ), ] _UpperCAmelCase : int = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str ) ->Any: '''simple docstring''' _UpperCAmelCase : int = hidden_state for layer_module in self.layers: _UpperCAmelCase : Tuple = layer_module(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : Tuple = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : str = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer _UpperCAmelCase : List[str] = [ # downsampling is done in the first layer with stride of 2 layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , name="layers.0" ), *[layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , name=F"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] ) ->List[str]: '''simple docstring''' for layer_module in self.layers: _UpperCAmelCase : Optional[int] = layer_module(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : int ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) _UpperCAmelCase : Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCamelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , depth=lowerCamelCase__ , name=F"""stages.{i+1}""" ) ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = True ) ->TFBaseModelOutputWithNoAttention: '''simple docstring''' _UpperCAmelCase : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase : Optional[Any] = hidden_states + (hidden_state,) _UpperCAmelCase : Dict = stage_module(lowerCamelCase__ ) if output_hidden_states: _UpperCAmelCase : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase__ , hidden_states=lowerCamelCase__ ) @keras_serializable class lowerCAmelCase__ ( tf.keras.layers.Layer ): lowerCAmelCase : Optional[Any] = RegNetConfig def __init__( self : Union[str, Any] , lowerCamelCase__ : Any , **lowerCamelCase__ : str ) ->int: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = config _UpperCAmelCase : Union[str, Any] = TFRegNetEmbeddings(lowerCamelCase__ , name="embedder" ) _UpperCAmelCase : Union[str, Any] = TFRegNetEncoder(lowerCamelCase__ , name="encoder" ) _UpperCAmelCase : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) @unpack_inputs def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : bool = False , ) ->TFBaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' _UpperCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.embedder(lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : str = self.encoder( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : Dict = encoder_outputs[0] _UpperCAmelCase : Dict = self.pooler(lowerCamelCase__ ) # Change to NCHW output format have uniformity in the modules _UpperCAmelCase : Union[str, Any] = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) _UpperCAmelCase : Tuple = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _UpperCAmelCase : List[str] = tuple([tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase__ , pooler_output=lowerCamelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = RegNetConfig lowerCAmelCase : Tuple = "regnet" lowerCAmelCase : Union[str, Any] = "pixel_values" @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} lowerCamelCase__ = r'\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCamelCase__ = r'\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : Any , **lowerCamelCase__ : List[str] ) ->Optional[int]: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Any=False , ) ->Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( pixel_values=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : List[Any] , **lowerCamelCase__ : Union[str, Any] ) ->Any: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = config.num_labels _UpperCAmelCase : Dict = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) # classification head _UpperCAmelCase : str = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict=False , ) ->Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : str = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : int = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase : Dict = self.classifier[0](lowerCamelCase__ ) _UpperCAmelCase : str = self.classifier[1](lowerCamelCase__ ) _UpperCAmelCase : Tuple = None if labels is None else self.hf_compute_loss(labels=lowerCamelCase__ , logits=lowerCamelCase__ ) if not return_dict: _UpperCAmelCase : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCamelCase__ , logits=lowerCamelCase__ , hidden_states=outputs.hidden_states )
322
0
'''simple docstring''' from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Optional[Any] , lowerCamelCase__ : NestedDataStructureLike[PathLike] , lowerCamelCase__ : Optional[NamedSplit] = None , lowerCamelCase__ : Optional[Features] = None , lowerCamelCase__ : str = None , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = False , lowerCamelCase__ : Optional[int] = None , **lowerCamelCase__ : Union[str, Any] , ) ->Union[str, Any]: '''simple docstring''' super().__init__( lowerCamelCase__ , split=lowerCamelCase__ , features=lowerCamelCase__ , cache_dir=lowerCamelCase__ , keep_in_memory=lowerCamelCase__ , streaming=lowerCamelCase__ , num_proc=lowerCamelCase__ , **lowerCamelCase__ , ) _UpperCAmelCase : List[str] = path_or_paths if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else {self.split: path_or_paths} _UpperCAmelCase : Optional[Any] = Text( cache_dir=lowerCamelCase__ , data_files=lowerCamelCase__ , features=lowerCamelCase__ , **lowerCamelCase__ , ) def lowerCAmelCase__ ( self : int ) ->Union[str, Any]: '''simple docstring''' if self.streaming: _UpperCAmelCase : Tuple = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: _UpperCAmelCase : Any = None _UpperCAmelCase : List[str] = None _UpperCAmelCase : List[Any] = None _UpperCAmelCase : List[Any] = None self.builder.download_and_prepare( download_config=lowerCamelCase__ , download_mode=lowerCamelCase__ , verification_mode=lowerCamelCase__ , base_path=lowerCamelCase__ , num_proc=self.num_proc , ) _UpperCAmelCase : Union[str, Any] = self.builder.as_dataset( split=self.split , verification_mode=lowerCamelCase__ , in_memory=self.keep_in_memory ) return dataset
359
'''simple docstring''' import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def __lowerCAmelCase (__lowerCAmelCase ): if is_torch_version("<" , "2.0.0" ) or not hasattr(__lowerCAmelCase , "_dynamo" ): return False return isinstance(__lowerCAmelCase , torch._dynamo.eval_frame.OptimizedModule ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase = True ): _UpperCAmelCase : Any = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) _UpperCAmelCase : Dict = is_compiled_module(__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : Optional[int] = model _UpperCAmelCase : Any = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = model.module if not keep_fpaa_wrapper: _UpperCAmelCase : List[Any] = getattr(__lowerCAmelCase , "forward" ) _UpperCAmelCase : Dict = model.__dict__.pop("_original_forward" , __lowerCAmelCase ) if original_forward is not None: while hasattr(__lowerCAmelCase , "__wrapped__" ): _UpperCAmelCase : Optional[int] = forward.__wrapped__ if forward == original_forward: break _UpperCAmelCase : Dict = forward if getattr(__lowerCAmelCase , "_converted_to_transformer_engine" , __lowerCAmelCase ): convert_model(__lowerCAmelCase , to_transformer_engine=__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : int = model _UpperCAmelCase : str = compiled_model return model def __lowerCAmelCase (): PartialState().wait_for_everyone() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(__lowerCAmelCase , __lowerCAmelCase ) elif PartialState().local_process_index == 0: torch.save(__lowerCAmelCase , __lowerCAmelCase ) @contextmanager def __lowerCAmelCase (**__lowerCAmelCase ): for key, value in kwargs.items(): _UpperCAmelCase : str = str(__lowerCAmelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def __lowerCAmelCase (__lowerCAmelCase ): if not hasattr(__lowerCAmelCase , "__qualname__" ) and not hasattr(__lowerCAmelCase , "__name__" ): _UpperCAmelCase : List[str] = getattr(__lowerCAmelCase , "__class__" , __lowerCAmelCase ) if hasattr(__lowerCAmelCase , "__qualname__" ): return obj.__qualname__ if hasattr(__lowerCAmelCase , "__name__" ): return obj.__name__ return str(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): for key, value in source.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = destination.setdefault(__lowerCAmelCase , {} ) merge_dicts(__lowerCAmelCase , __lowerCAmelCase ) else: _UpperCAmelCase : Optional[int] = value return destination def __lowerCAmelCase (__lowerCAmelCase = None ): if port is None: _UpperCAmelCase : Tuple = 29_500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
322
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable lowerCamelCase__ = list[list[float | int]] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(size + 1 )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : float for row in range(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = matrix[row][col] _UpperCAmelCase : Optional[int] = vector[row][0] _UpperCAmelCase : int = 0 _UpperCAmelCase : Union[str, Any] = 0 while row < size and col < size: # pivoting _UpperCAmelCase : Optional[Any] = max((abs(augmented[rowa][col] ), rowa) for rowa in range(__lowerCAmelCase , __lowerCAmelCase ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: _UpperCAmelCase : str = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , __lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[rowa][col] / augmented[row][col] _UpperCAmelCase : Optional[Any] = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , __lowerCAmelCase ): for row in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[row][col] / augmented[col][col] for cola in range(__lowerCAmelCase , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(__lowerCAmelCase ) ] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(__lowerCAmelCase )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix = [[0] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int for x_val, y_val in enumerate(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = (x_val + 1) ** (size - col - 1) _UpperCAmelCase : int = y_val _UpperCAmelCase : List[str] = solve(__lowerCAmelCase , __lowerCAmelCase ) def interpolated_func(__lowerCAmelCase ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(__lowerCAmelCase ) ) return interpolated_func def __lowerCAmelCase (__lowerCAmelCase ): return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def __lowerCAmelCase (__lowerCAmelCase = question_function , __lowerCAmelCase = 10 ): _UpperCAmelCase : list[int] = [func(__lowerCAmelCase ) for x_val in range(1 , order + 1 )] _UpperCAmelCase : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] _UpperCAmelCase : int = 0 _UpperCAmelCase : Callable[[int], int] _UpperCAmelCase : int for poly in polynomials: _UpperCAmelCase : int = 1 while func(__lowerCAmelCase ) == poly(__lowerCAmelCase ): x_val += 1 ret += poly(__lowerCAmelCase ) return ret if __name__ == "__main__": print(F'''{solution() = }''')
360
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): if number < 0: raise ValueError("number must not be negative" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = min(__lowerCAmelCase ) # min() finds the minimum value _UpperCAmelCase : Any = max(__lowerCAmelCase ) # max() finds the maximum value _UpperCAmelCase : Optional[Any] = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size _UpperCAmelCase : Optional[Any] = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(__lowerCAmelCase , __lowerCAmelCase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. _UpperCAmelCase : Any = 0 for count in range(__lowerCAmelCase ): while holes[count] > 0: holes[count] -= 1 _UpperCAmelCase : str = count + min_val i += 1 def __lowerCAmelCase (): _UpperCAmelCase : Tuple = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(__lowerCAmelCase ) print("Sorted order is:" , " ".join(__lowerCAmelCase ) ) if __name__ == "__main__": main()
361
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('Program to check whether a number is a Perfect number or not...') lowerCamelCase__ = int(input('Enter number: ').strip()) print(F'''{number} is {"" if perfect(number) else "not "}a Perfect Number.''')
322
0
'''simple docstring''' import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCamelCase__ = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class lowerCAmelCase__ ( unittest.TestCase ): @classmethod def lowerCAmelCase__ ( cls : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : Tuple = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def lowerCAmelCase__ ( cls : Union[str, Any] ) ->int: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def lowerCAmelCase__ ( self : int ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ , repo_id="test-config" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : Dict = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ , repo_id="valid_org/test-config-org" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : int = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' CustomConfig.register_for_auto_class() _UpperCAmelCase : int = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) _UpperCAmelCase : str = AutoConfig.from_pretrained(F"""{USER}/test-dynamic-config""" , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : List[str] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated _UpperCAmelCase : Any = c.n_embd + 1 # int _UpperCAmelCase : List[Any] = c.resid_pdrop + 1.0 # float _UpperCAmelCase : Tuple = not c.scale_attn_weights # bool _UpperCAmelCase : List[Any] = c.summary_type + "foo" # str c.update_from_string( F"""n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}""" ) self.assertEqual(lowerCamelCase__ , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(lowerCamelCase__ , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(lowerCamelCase__ , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(lowerCamelCase__ , c.summary_type , "mismatch for key: summary_type" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = PretrainedConfig() _UpperCAmelCase : Tuple = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase__ , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) _UpperCAmelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase__ , lowerCamelCase__ )] if len(lowerCamelCase__ ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" F""" {', '.join(lowerCamelCase__ )}.""" ) def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): # config is in subfolder, the following should not work without specifying the subfolder _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = mock.Mock() _UpperCAmelCase : List[str] = 5_00 _UpperCAmelCase : Dict = {} _UpperCAmelCase : Tuple = HTTPError _UpperCAmelCase : Any = {} # Download this model to make sure it's in the cache. _UpperCAmelCase : int = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=lowerCamelCase__ ) as mock_head: _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = AutoConfig.from_pretrained("bert-base-cased" ) _UpperCAmelCase : str = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 _UpperCAmelCase : Optional[int] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 _UpperCAmelCase : Dict = ["config.42.0.0.json"] _UpperCAmelCase : Union[str, Any] = 7_68 configuration.save_pretrained(lowerCamelCase__ ) shutil.move(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , os.path.join(lowerCamelCase__ , "config.42.0.0.json" ) ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = "hf-internal-testing/test-two-configs" import transformers as new_transformers _UpperCAmelCase : Any = "v4.0.0" _UpperCAmelCase : Optional[Any] = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers _UpperCAmelCase : List[Any] = "v3.0.0" _UpperCAmelCase : int = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(old_configuration.hidden_size , 7_68 )
362
'''simple docstring''' from collections.abc import Sequence def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return sum(c * (x**i) for i, c in enumerate(__lowerCAmelCase ) ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Dict = 0.0 for coeff in reversed(__lowerCAmelCase ): _UpperCAmelCase : int = result * x + coeff return result if __name__ == "__main__": lowerCamelCase__ = (0.0, 0.0, 5.0, 9.3, 7.0) lowerCamelCase__ = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
322
0
'''simple docstring''' from collections.abc import Sequence def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): return sum(c * (x**i) for i, c in enumerate(__lowerCAmelCase ) ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Dict = 0.0 for coeff in reversed(__lowerCAmelCase ): _UpperCAmelCase : int = result * x + coeff return result if __name__ == "__main__": lowerCamelCase__ = (0.0, 0.0, 5.0, 9.3, 7.0) lowerCamelCase__ = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
363
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : List[Any] = len(__lowerCAmelCase ) _UpperCAmelCase : Tuple = sum(__lowerCAmelCase ) _UpperCAmelCase : List[Any] = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _UpperCAmelCase : Any = True for i in range(1 , s + 1 ): _UpperCAmelCase : List[Any] = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): _UpperCAmelCase : Optional[int] = dp[i][j - 1] if arr[i - 1] <= j: _UpperCAmelCase : Any = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: _UpperCAmelCase : List[Any] = s - 2 * j break return diff
322
0
'''simple docstring''' lowerCamelCase__ = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = set() # keep track of all the paths to be checked _UpperCAmelCase : int = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue _UpperCAmelCase : Union[str, Any] = queue.pop(0 ) # get the last node from the path _UpperCAmelCase : Tuple = path[-1] if node not in explored: _UpperCAmelCase : List[str] = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: _UpperCAmelCase : Optional[Any] = list(__lowerCAmelCase ) new_path.append(__lowerCAmelCase ) queue.append(__lowerCAmelCase ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(__lowerCAmelCase ) # in case there's no path between the 2 nodes return [] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 _UpperCAmelCase : Union[str, Any] = [start] _UpperCAmelCase : Dict = set(__lowerCAmelCase ) # Keep tab on distances from `start` node. _UpperCAmelCase : Optional[int] = {start: 0, target: -1} while queue: _UpperCAmelCase : Any = queue.pop(0 ) if node == target: _UpperCAmelCase : int = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(__lowerCAmelCase ) queue.append(__lowerCAmelCase ) _UpperCAmelCase : Dict = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, 'G', 'D')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, 'G', 'D')) # returns 4
364
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'microsoft/resnet-50': 'https://huggingface.co/microsoft/resnet-50/blob/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): lowerCAmelCase : int = "resnet" lowerCAmelCase : Union[str, Any] = ["basic", "bottleneck"] def __init__( self : Dict , lowerCamelCase__ : Tuple=3 , lowerCamelCase__ : Any=64 , lowerCamelCase__ : Optional[int]=[2_56, 5_12, 10_24, 20_48] , lowerCamelCase__ : int=[3, 4, 6, 3] , lowerCamelCase__ : Dict="bottleneck" , lowerCamelCase__ : Dict="relu" , lowerCamelCase__ : List[Any]=False , lowerCamelCase__ : Any=None , lowerCamelCase__ : int=None , **lowerCamelCase__ : Tuple , ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) if layer_type not in self.layer_types: raise ValueError(F"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) _UpperCAmelCase : str = num_channels _UpperCAmelCase : List[str] = embedding_size _UpperCAmelCase : Tuple = hidden_sizes _UpperCAmelCase : Dict = depths _UpperCAmelCase : List[Any] = layer_type _UpperCAmelCase : Optional[int] = hidden_act _UpperCAmelCase : Tuple = downsample_in_first_stage _UpperCAmelCase : str = ["stem"] + [F"""stage{idx}""" for idx in range(1 , len(lowerCamelCase__ ) + 1 )] _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = get_aligned_output_features_output_indices( out_features=lowerCamelCase__ , out_indices=lowerCamelCase__ , stage_names=self.stage_names ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Optional[Any] = version.parse("1.11" ) @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowerCAmelCase__ ( self : str ) ->float: '''simple docstring''' return 1E-3
322
0
'''simple docstring''' import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def __lowerCAmelCase (__lowerCAmelCase ): if isinstance(__lowerCAmelCase , collections.abc.Iterable ): return x return (x, x) @require_flax class lowerCAmelCase__ : def lowerCAmelCase__ ( self : str , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Tuple ) ->Optional[int]: '''simple docstring''' pass def lowerCAmelCase__ ( self : Optional[int] ) ->Tuple: '''simple docstring''' pass def lowerCAmelCase__ ( self : Dict ) ->List[Any]: '''simple docstring''' pass def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : np.ndarray , lowerCamelCase__ : float ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : int = np.abs((a - b) ).max() self.assertLessEqual(lowerCamelCase__ , lowerCamelCase__ , F"""Difference between torch and flax is {diff} (>= {tol}).""" ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : int , lowerCamelCase__ : Dict , lowerCamelCase__ : Tuple , lowerCamelCase__ : List[Any]=None , **lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = VisionTextDualEncoderConfig.from_vision_text_configs(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel(lowerCamelCase__ ) _UpperCAmelCase : str = model(input_ids=lowerCamelCase__ , pixel_values=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim) ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : str=None , **lowerCamelCase__ : List[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.get_vision_text_model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = {"vision_model": vision_model, "text_model": text_model} _UpperCAmelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = model(input_ids=lowerCamelCase__ , pixel_values=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Optional[Any]=None , **lowerCamelCase__ : Optional[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.get_vision_text_model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Any = {"vision_model": vision_model, "text_model": text_model} _UpperCAmelCase : List[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCamelCase__ ) _UpperCAmelCase : Tuple = model(input_ids=lowerCamelCase__ , pixel_values=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = FlaxVisionTextDualEncoderModel.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : List[str] = model(input_ids=lowerCamelCase__ , pixel_values=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) _UpperCAmelCase : Any = after_output[0] _UpperCAmelCase : Union[str, Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Tuple , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : str , lowerCamelCase__ : List[str] , lowerCamelCase__ : List[str]=None , **lowerCamelCase__ : Dict ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.get_vision_text_model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[str] = {"vision_model": vision_model, "text_model": text_model} _UpperCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**lowerCamelCase__ ) _UpperCAmelCase : str = model( input_ids=lowerCamelCase__ , pixel_values=lowerCamelCase__ , attention_mask=lowerCamelCase__ , output_attentions=lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = output.vision_model_output.attentions self.assertEqual(len(lowerCamelCase__ ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) _UpperCAmelCase : Dict = to_atuple(vision_model.config.image_size ) _UpperCAmelCase : str = to_atuple(vision_model.config.patch_size ) _UpperCAmelCase : Dict = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _UpperCAmelCase : int = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _UpperCAmelCase : Optional[int] = output.text_model_output.attentions self.assertEqual(len(lowerCamelCase__ ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple , lowerCamelCase__ : int ) ->Optional[int]: '''simple docstring''' pt_model.to(lowerCamelCase__ ) pt_model.eval() # prepare inputs _UpperCAmelCase : Any = inputs_dict _UpperCAmelCase : Tuple = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): _UpperCAmelCase : Optional[int] = pt_model(**lowerCamelCase__ ).to_tuple() _UpperCAmelCase : str = fx_model(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(lowerCamelCase__ , pt_output.numpy() , 4E-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : int = FlaxVisionTextDualEncoderModel.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) _UpperCAmelCase : Any = fx_model_loaded(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) , "Output lengths differ between Flax and PyTorch" ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(lowerCamelCase__ , pt_output.numpy() , 4E-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Tuple = VisionTextDualEncoderModel.from_pretrained(lowerCamelCase__ , from_flax=lowerCamelCase__ ) pt_model_loaded.to(lowerCamelCase__ ) pt_model_loaded.eval() with torch.no_grad(): _UpperCAmelCase : int = pt_model_loaded(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) , "Output lengths differ between Flax and PyTorch" ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(lowerCamelCase__ , pt_output_loaded.numpy() , 4E-2 ) def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Any , lowerCamelCase__ : int , lowerCamelCase__ : Tuple ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = VisionTextDualEncoderModel(lowerCamelCase__ ) _UpperCAmelCase : Tuple = FlaxVisionTextDualEncoderModel(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = fx_state self.check_pt_flax_equivalence(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Optional[Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[int] = VisionTextDualEncoderConfig.from_vision_text_configs(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : int = VisionTextDualEncoderModel(lowerCamelCase__ ) _UpperCAmelCase : List[str] = FlaxVisionTextDualEncoderModel(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = load_flax_weights_in_pytorch_model(lowerCamelCase__ , fx_model.params ) self.check_pt_flax_equivalence(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : int = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**lowerCamelCase__ ) def lowerCAmelCase__ ( self : str ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[str] ) ->int: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.prepare_config_and_inputs() self.check_save_load(**lowerCamelCase__ ) def lowerCAmelCase__ ( self : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**lowerCamelCase__ ) @is_pt_flax_cross_test def lowerCAmelCase__ ( self : Union[str, Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.prepare_config_and_inputs() _UpperCAmelCase : Optional[int] = config_inputs_dict.pop("vision_config" ) _UpperCAmelCase : Dict = config_inputs_dict.pop("text_config" ) _UpperCAmelCase : Dict = config_inputs_dict self.check_equivalence_pt_to_flax(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) self.check_equivalence_flax_to_pt(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) @slow def lowerCAmelCase__ ( self : str ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Dict = self.get_pretrained_model_and_inputs() _UpperCAmelCase : Dict = model_a(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : int = FlaxVisionTextDualEncoderModel.from_pretrained(lowerCamelCase__ ) _UpperCAmelCase : str = model_a(**lowerCamelCase__ ) _UpperCAmelCase : int = after_outputs[0] _UpperCAmelCase : Any = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1E-5 ) @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): def lowerCAmelCase__ ( self : Optional[Any] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Any = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-bert" , vision_from_pt=lowerCamelCase__ , text_from_pt=lowerCamelCase__ , ) _UpperCAmelCase : Optional[int] = 13 _UpperCAmelCase : int = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _UpperCAmelCase : Dict = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _UpperCAmelCase : Dict = random_attention_mask([batch_size, 4] ) _UpperCAmelCase : Optional[int] = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def lowerCAmelCase__ ( self : str , lowerCamelCase__ : str , lowerCamelCase__ : str ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : List[str] = FlaxViTModel(lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = FlaxBertModel(lowerCamelCase__ ) return vision_model, text_model def lowerCAmelCase__ ( self : Dict ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = FlaxViTModelTester(self ) _UpperCAmelCase : List[str] = FlaxBertModelTester(self ) _UpperCAmelCase : Any = vit_model_tester.prepare_config_and_inputs() _UpperCAmelCase : int = bert_model_tester.prepare_config_and_inputs() _UpperCAmelCase : List[Any] = vision_config_and_inputs _UpperCAmelCase : List[str] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): def lowerCAmelCase__ ( self : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : Tuple = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-clip" , "hf-internal-testing/tiny-bert" , vision_from_pt=lowerCamelCase__ , text_from_pt=lowerCamelCase__ , ) _UpperCAmelCase : List[str] = 13 _UpperCAmelCase : Optional[int] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) _UpperCAmelCase : Dict = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) _UpperCAmelCase : str = random_attention_mask([batch_size, 4] ) _UpperCAmelCase : Dict = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Optional[int] , lowerCamelCase__ : Optional[int] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = FlaxCLIPVisionModel(lowerCamelCase__ ) _UpperCAmelCase : Any = FlaxBertModel(lowerCamelCase__ ) return vision_model, text_model def lowerCAmelCase__ ( self : List[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = FlaxCLIPVisionModelTester(self ) _UpperCAmelCase : Union[str, Any] = FlaxBertModelTester(self ) _UpperCAmelCase : str = clip_model_tester.prepare_config_and_inputs() _UpperCAmelCase : int = bert_model_tester.prepare_config_and_inputs() _UpperCAmelCase : List[str] = vision_config_and_inputs _UpperCAmelCase : int = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Dict ) ->Dict: '''simple docstring''' _UpperCAmelCase : Optional[int] = FlaxVisionTextDualEncoderModel.from_pretrained("clip-italian/clip-italian" , logit_scale_init_value=1.0 ) _UpperCAmelCase : Union[str, Any] = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian" ) _UpperCAmelCase : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) _UpperCAmelCase : Optional[int] = processor( text=["una foto di un gatto", "una foto di un cane"] , images=lowerCamelCase__ , padding=lowerCamelCase__ , return_tensors="np" ) _UpperCAmelCase : Dict = model(**lowerCamelCase__ ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) _UpperCAmelCase : Optional[int] = np.array([[1.2_2_8_4_7_2_7, 0.3_1_0_4_1_2_2]] ) self.assertTrue(np.allclose(outputs.logits_per_image , lowerCamelCase__ , atol=1E-3 ) )
365
'''simple docstring''' from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowerCamelCase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __lowerCAmelCase (__lowerCAmelCase ): if isinstance(__lowerCAmelCase , torch.Tensor ): return image elif isinstance(__lowerCAmelCase , PIL.Image.Image ): _UpperCAmelCase : int = [image] _UpperCAmelCase : str = [trans(img.convert("RGB" ) ) for img in image] _UpperCAmelCase : Optional[Any] = torch.stack(__lowerCAmelCase ) return image class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Tuple , lowerCamelCase__ : int , lowerCamelCase__ : int ) ->int: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM _UpperCAmelCase : Tuple = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : str ) ->Union[str, Any]: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(F"""The value of strength should in [0.0, 1.0] but is {strength}""" ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Dict , lowerCamelCase__ : List[str] , lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : str = min(int(num_inference_steps * strength ) , lowerCamelCase__ ) _UpperCAmelCase : str = max(num_inference_steps - init_timestep , 0 ) _UpperCAmelCase : List[str] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Any , lowerCamelCase__ : str , lowerCamelCase__ : str , lowerCamelCase__ : Dict , lowerCamelCase__ : Optional[Any]=None ) ->str: '''simple docstring''' if not isinstance(lowerCamelCase__ , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowerCamelCase__ )}""" ) _UpperCAmelCase : Union[str, Any] = image.to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(lowerCamelCase__ )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) _UpperCAmelCase : List[str] = init_latents.shape _UpperCAmelCase : Optional[int] = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=lowerCamelCase__ , dtype=lowerCamelCase__ ) # get latents print("add noise to latents at timestep" , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.scheduler.add_noise(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = init_latents return latents @torch.no_grad() def __call__( self : Any , lowerCamelCase__ : Union[torch.FloatTensor, PIL.Image.Image] = None , lowerCamelCase__ : float = 0.8 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase__ : float = 0.0 , lowerCamelCase__ : int = 50 , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , ) ->Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowerCamelCase__ ) # 2. Preprocess image _UpperCAmelCase : Dict = preprocess(lowerCamelCase__ ) # 3. set timesteps self.scheduler.set_timesteps(lowerCamelCase__ , device=self.device ) _UpperCAmelCase , _UpperCAmelCase : Any = self.get_timesteps(lowerCamelCase__ , lowerCamelCase__ , self.device ) _UpperCAmelCase : List[Any] = timesteps[:1].repeat(lowerCamelCase__ ) # 4. Prepare latent variables _UpperCAmelCase : Optional[int] = self.prepare_latents(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , self.unet.dtype , self.device , lowerCamelCase__ ) _UpperCAmelCase : Any = latents # 5. Denoising loop for t in self.progress_bar(lowerCamelCase__ ): # 1. predict noise model_output _UpperCAmelCase : Union[str, Any] = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _UpperCAmelCase : int = self.scheduler.step( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , eta=lowerCamelCase__ , use_clipped_model_output=lowerCamelCase__ , generator=lowerCamelCase__ , ).prev_sample _UpperCAmelCase : Dict = (image / 2 + 0.5).clamp(0 , 1 ) _UpperCAmelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCAmelCase : str = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowerCamelCase__ )
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = len(__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): _UpperCAmelCase : List[Any] = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): _UpperCAmelCase : int = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: _UpperCAmelCase : str = subset[i - 1][j] if arr[i - 1] <= j: _UpperCAmelCase : Any = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
366
'''simple docstring''' from __future__ import annotations from collections.abc import Callable lowerCamelCase__ = list[list[float | int]] def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(size + 1 )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : float for row in range(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = matrix[row][col] _UpperCAmelCase : Optional[int] = vector[row][0] _UpperCAmelCase : int = 0 _UpperCAmelCase : Union[str, Any] = 0 while row < size and col < size: # pivoting _UpperCAmelCase : Optional[Any] = max((abs(augmented[rowa][col] ), rowa) for rowa in range(__lowerCAmelCase , __lowerCAmelCase ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: _UpperCAmelCase , _UpperCAmelCase : str = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , __lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[rowa][col] / augmented[row][col] _UpperCAmelCase : Optional[Any] = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , __lowerCAmelCase ): for row in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = augmented[row][col] / augmented[col][col] for cola in range(__lowerCAmelCase , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(__lowerCAmelCase ) ] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : int = len(__lowerCAmelCase ) _UpperCAmelCase : Matrix = [[0 for _ in range(__lowerCAmelCase )] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix = [[0] for _ in range(__lowerCAmelCase )] _UpperCAmelCase : Matrix _UpperCAmelCase : int _UpperCAmelCase : int _UpperCAmelCase : int for x_val, y_val in enumerate(__lowerCAmelCase ): for col in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = (x_val + 1) ** (size - col - 1) _UpperCAmelCase : int = y_val _UpperCAmelCase : List[str] = solve(__lowerCAmelCase , __lowerCAmelCase ) def interpolated_func(__lowerCAmelCase ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(__lowerCAmelCase ) ) return interpolated_func def __lowerCAmelCase (__lowerCAmelCase ): return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def __lowerCAmelCase (__lowerCAmelCase = question_function , __lowerCAmelCase = 10 ): _UpperCAmelCase : list[int] = [func(__lowerCAmelCase ) for x_val in range(1 , order + 1 )] _UpperCAmelCase : list[Callable[[int], int]] = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] _UpperCAmelCase : int = 0 _UpperCAmelCase : Callable[[int], int] _UpperCAmelCase : int for poly in polynomials: _UpperCAmelCase : int = 1 while func(__lowerCAmelCase ) == poly(__lowerCAmelCase ): x_val += 1 ret += poly(__lowerCAmelCase ) return ret if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' from __future__ import annotations import queue class lowerCAmelCase__ : def __init__( self : List[str] , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = data _UpperCAmelCase : str = None _UpperCAmelCase : Union[str, Any] = None def __lowerCAmelCase (): print("\n********Press N to stop entering at any point of time********\n" ) _UpperCAmelCase : str = input("Enter the value of the root node: " ).strip().lower() _UpperCAmelCase : queue.Queue = queue.Queue() _UpperCAmelCase : Tuple = TreeNode(int(__lowerCAmelCase ) ) q.put(__lowerCAmelCase ) while not q.empty(): _UpperCAmelCase : str = q.get() _UpperCAmelCase : Tuple = F"""Enter the left node of {node_found.data}: """ _UpperCAmelCase : List[Any] = input(__lowerCAmelCase ).strip().lower() or "n" if check == "n": return tree_node _UpperCAmelCase : Dict = TreeNode(int(__lowerCAmelCase ) ) _UpperCAmelCase : int = left_node q.put(__lowerCAmelCase ) _UpperCAmelCase : List[str] = F"""Enter the right node of {node_found.data}: """ _UpperCAmelCase : Any = input(__lowerCAmelCase ).strip().lower() or "n" if check == "n": return tree_node _UpperCAmelCase : Optional[int] = TreeNode(int(__lowerCAmelCase ) ) _UpperCAmelCase : Optional[Any] = right_node q.put(__lowerCAmelCase ) raise def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return print(node.data , end="," ) pre_order(node.left ) pre_order(node.right ) def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return in_order(node.left ) print(node.data , end="," ) in_order(node.right ) def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end="," ) def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return _UpperCAmelCase : queue.Queue = queue.Queue() q.put(__lowerCAmelCase ) while not q.empty(): _UpperCAmelCase : Any = q.get() print(node_dequeued.data , end="," ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return _UpperCAmelCase : queue.Queue = queue.Queue() q.put(__lowerCAmelCase ) while not q.empty(): _UpperCAmelCase : Dict = [] while not q.empty(): _UpperCAmelCase : Optional[Any] = q.get() print(node_dequeued.data , end="," ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return _UpperCAmelCase : list[TreeNode] = [] _UpperCAmelCase : Tuple = node while n or stack: while n: # start from root node, find its left child print(n.data , end="," ) stack.append(__lowerCAmelCase ) _UpperCAmelCase : int = n.left # end of while means current node doesn't have left child _UpperCAmelCase : Any = stack.pop() # start to traverse its right child _UpperCAmelCase : str = n.right def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return _UpperCAmelCase : list[TreeNode] = [] _UpperCAmelCase : Any = node while n or stack: while n: stack.append(__lowerCAmelCase ) _UpperCAmelCase : Tuple = n.left _UpperCAmelCase : Dict = stack.pop() print(n.data , end="," ) _UpperCAmelCase : Optional[Any] = n.right def __lowerCAmelCase (__lowerCAmelCase ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return _UpperCAmelCase : Tuple = [], [] _UpperCAmelCase : str = node stacka.append(__lowerCAmelCase ) while stacka: # to find the reversed order of post order, store it in stack2 _UpperCAmelCase : Optional[Any] = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(__lowerCAmelCase ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end="," ) def __lowerCAmelCase (__lowerCAmelCase = "" , __lowerCAmelCase=50 , __lowerCAmelCase="*" ): if not s: return "\n" + width * char _UpperCAmelCase : Optional[Any] = divmod(width - len(__lowerCAmelCase ) - 2 , 2 ) return F"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt('Binary Tree Traversals')) lowerCamelCase__ = build_tree() print(prompt('Pre Order Traversal')) pre_order(node) print(prompt() + '\n') print(prompt('In Order Traversal')) in_order(node) print(prompt() + '\n') print(prompt('Post Order Traversal')) post_order(node) print(prompt() + '\n') print(prompt('Level Order Traversal')) level_order(node) print(prompt() + '\n') print(prompt('Actual Level Order Traversal')) level_order_actual(node) print('*' * 50 + '\n') print(prompt('Pre Order Traversal - Iteration Version')) pre_order_iter(node) print(prompt() + '\n') print(prompt('In Order Traversal - Iteration Version')) in_order_iter(node) print(prompt() + '\n') print(prompt('Post Order Traversal - Iteration Version')) post_order_iter(node) print(prompt())
367
'''simple docstring''' from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
322
0
'''simple docstring''' from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=UpperCAmelCase__ ) class lowerCAmelCase__ ( UpperCAmelCase__ ): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCAmelCase : str = field(default="summarization" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCAmelCase : ClassVar[Features] = Features({"text": Value("string" )} ) lowerCAmelCase : ClassVar[Features] = Features({"summary": Value("string" )} ) lowerCAmelCase : str = "text" lowerCAmelCase : str = "summary" @property def lowerCAmelCase__ ( self : Tuple ) ->Dict[str, str]: '''simple docstring''' return {self.text_column: "text", self.summary_column: "summary"}
368
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar lowerCamelCase__ = TypeVar('T') class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Union[str, Any] , lowerCamelCase__ : T ) ->Tuple: '''simple docstring''' _UpperCAmelCase : Dict = data _UpperCAmelCase : Node[T] | None = None def __str__( self : Any ) ->str: '''simple docstring''' return F"""{self.data}""" class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Tuple ) ->None: '''simple docstring''' _UpperCAmelCase : Node[T] | None = None def __iter__( self : List[str] ) ->Iterator[T]: '''simple docstring''' _UpperCAmelCase : Any = self.top while node: yield node.data _UpperCAmelCase : Dict = node.next def __str__( self : Dict ) ->str: '''simple docstring''' return "->".join([str(lowerCamelCase__ ) for item in self] ) def __len__( self : Optional[int] ) ->int: '''simple docstring''' return len(tuple(iter(self ) ) ) def lowerCAmelCase__ ( self : List[Any] ) ->bool: '''simple docstring''' return self.top is None def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : T ) ->None: '''simple docstring''' _UpperCAmelCase : List[Any] = Node(lowerCamelCase__ ) if not self.is_empty(): _UpperCAmelCase : Tuple = self.top _UpperCAmelCase : List[str] = node def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("pop from empty stack" ) assert isinstance(self.top , lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.top _UpperCAmelCase : Optional[Any] = self.top.next return pop_node.data def lowerCAmelCase__ ( self : Union[str, Any] ) ->T: '''simple docstring''' if self.is_empty(): raise IndexError("peek from empty stack" ) assert self.top is not None return self.top.data def lowerCAmelCase__ ( self : List[Any] ) ->None: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = None if __name__ == "__main__": from doctest import testmod testmod()
322
0
'''simple docstring''' import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Any = (DPMSolverSinglestepScheduler,) lowerCAmelCase : Any = (("num_inference_steps", 25),) def lowerCAmelCase__ ( self : Optional[Any] , **lowerCamelCase__ : Any ) ->Tuple: '''simple docstring''' _UpperCAmelCase : List[Any] = { "num_train_timesteps": 10_00, "beta_start": 0.0_0_0_1, "beta_end": 0.0_2, "beta_schedule": "linear", "solver_order": 2, "prediction_type": "epsilon", "thresholding": False, "sample_max_value": 1.0, "algorithm_type": "dpmsolver++", "solver_type": "midpoint", "lambda_min_clipped": -float("inf" ), "variance_type": None, } config.update(**lowerCamelCase__ ) return config def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : int=0 , **lowerCamelCase__ : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : List[str] = dict(self.forward_default_kwargs ) _UpperCAmelCase : List[str] = kwargs.pop("num_inference_steps" , lowerCamelCase__ ) _UpperCAmelCase : List[str] = self.dummy_sample _UpperCAmelCase : Optional[Any] = 0.1 * sample _UpperCAmelCase : str = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: _UpperCAmelCase : Dict = self.get_scheduler_config(**lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals _UpperCAmelCase : Union[str, Any] = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) _UpperCAmelCase : Dict = scheduler_class.from_pretrained(lowerCamelCase__ ) new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals _UpperCAmelCase : Any = dummy_past_residuals[: new_scheduler.config.solver_order] _UpperCAmelCase : Any = sample, sample for t in range(lowerCamelCase__ , time_step + scheduler.config.solver_order + 1 ): _UpperCAmelCase : Dict = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _UpperCAmelCase : Dict = new_scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' pass def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : int=0 , **lowerCamelCase__ : Any ) ->Dict: '''simple docstring''' _UpperCAmelCase : int = dict(self.forward_default_kwargs ) _UpperCAmelCase : List[Any] = kwargs.pop("num_inference_steps" , lowerCamelCase__ ) _UpperCAmelCase : Tuple = self.dummy_sample _UpperCAmelCase : List[Any] = 0.1 * sample _UpperCAmelCase : List[str] = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: _UpperCAmelCase : List[str] = self.get_scheduler_config() _UpperCAmelCase : List[str] = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _UpperCAmelCase : Optional[Any] = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) _UpperCAmelCase : str = scheduler_class.from_pretrained(lowerCamelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residual (must be after setting timesteps) _UpperCAmelCase : Dict = dummy_past_residuals[: new_scheduler.config.solver_order] _UpperCAmelCase : Any = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _UpperCAmelCase : str = new_scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : Optional[int]=None , **lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' if scheduler is None: _UpperCAmelCase : Optional[Any] = self.scheduler_classes[0] _UpperCAmelCase : Optional[Any] = self.get_scheduler_config(**lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = scheduler_class(**lowerCamelCase__ ) _UpperCAmelCase : Tuple = self.scheduler_classes[0] _UpperCAmelCase : int = self.get_scheduler_config(**lowerCamelCase__ ) _UpperCAmelCase : str = scheduler_class(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = 10 _UpperCAmelCase : List[str] = self.dummy_model() _UpperCAmelCase : Optional[int] = self.dummy_sample_deter scheduler.set_timesteps(lowerCamelCase__ ) for i, t in enumerate(scheduler.timesteps ): _UpperCAmelCase : Dict = model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Tuple = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample return sample def lowerCAmelCase__ ( self : int ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Dict = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) _UpperCAmelCase : str = 50 _UpperCAmelCase : int = self.dummy_model() _UpperCAmelCase : Optional[int] = self.dummy_sample_deter scheduler.set_timesteps(lowerCamelCase__ ) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:] ): _UpperCAmelCase : int = model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample _UpperCAmelCase : List[str] = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_5_7_4 ) < 1E-3 def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' for timesteps in [25, 50, 1_00, 9_99, 10_00]: self.check_over_configs(num_train_timesteps=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : str = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) _UpperCAmelCase : str = self.full_loop(scheduler=lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_7_9_1 ) < 1E-3 _UpperCAmelCase : Tuple = DEISMultistepScheduler.from_config(scheduler.config ) _UpperCAmelCase : Dict = DPMSolverMultistepScheduler.from_config(scheduler.config ) _UpperCAmelCase : List[str] = UniPCMultistepScheduler.from_config(scheduler.config ) _UpperCAmelCase : int = DPMSolverSinglestepScheduler.from_config(scheduler.config ) _UpperCAmelCase : Tuple = self.full_loop(scheduler=lowerCamelCase__ ) _UpperCAmelCase : Any = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_7_9_1 ) < 1E-3 def lowerCAmelCase__ ( self : Any ) ->Tuple: '''simple docstring''' self.check_over_configs(thresholding=lowerCamelCase__ ) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowerCamelCase__ , prediction_type=lowerCamelCase__ , sample_max_value=lowerCamelCase__ , algorithm_type="dpmsolver++" , solver_order=lowerCamelCase__ , solver_type=lowerCamelCase__ , ) def lowerCAmelCase__ ( self : int ) ->Union[str, Any]: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowerCamelCase__ , solver_type=lowerCamelCase__ , prediction_type=lowerCamelCase__ , algorithm_type=lowerCamelCase__ , ) _UpperCAmelCase : Union[str, Any] = self.full_loop( solver_order=lowerCamelCase__ , solver_type=lowerCamelCase__ , prediction_type=lowerCamelCase__ , algorithm_type=lowerCamelCase__ , ) assert not torch.isnan(lowerCamelCase__ ).any(), "Samples have nan numbers" def lowerCAmelCase__ ( self : Tuple ) ->int: '''simple docstring''' self.check_over_configs(lower_order_final=lowerCamelCase__ ) self.check_over_configs(lower_order_final=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Union[str, Any] ) ->Dict: '''simple docstring''' self.check_over_configs(lambda_min_clipped=-float("inf" ) ) self.check_over_configs(lambda_min_clipped=-5.1 ) def lowerCAmelCase__ ( self : int ) ->Union[str, Any]: '''simple docstring''' self.check_over_configs(variance_type=lowerCamelCase__ ) self.check_over_configs(variance_type="learned_range" ) def lowerCAmelCase__ ( self : Tuple ) ->str: '''simple docstring''' for num_inference_steps in [1, 2, 3, 5, 10, 50, 1_00, 9_99, 10_00]: self.check_over_forward(num_inference_steps=lowerCamelCase__ , time_step=0 ) def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.full_loop() _UpperCAmelCase : Union[str, Any] = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_7_9_1 ) < 1E-3 def lowerCAmelCase__ ( self : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : List[str] = self.full_loop(use_karras_sigmas=lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.2_2_4_8 ) < 1E-3 def lowerCAmelCase__ ( self : int ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : str = self.full_loop(prediction_type="v_prediction" ) _UpperCAmelCase : Tuple = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.1_4_5_3 ) < 1E-3 def lowerCAmelCase__ ( self : Tuple ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = self.full_loop(prediction_type="v_prediction" , use_karras_sigmas=lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_mean.item() - 0.0_6_4_9 ) < 1E-3 def lowerCAmelCase__ ( self : Any ) ->Tuple: '''simple docstring''' _UpperCAmelCase : List[str] = self.scheduler_classes[0] _UpperCAmelCase : Optional[Any] = self.get_scheduler_config(thresholding=lowerCamelCase__ , dynamic_thresholding_ratio=0 ) _UpperCAmelCase : Optional[int] = scheduler_class(**lowerCamelCase__ ) _UpperCAmelCase : int = 10 _UpperCAmelCase : Optional[Any] = self.dummy_model() _UpperCAmelCase : List[Any] = self.dummy_sample_deter.half() scheduler.set_timesteps(lowerCamelCase__ ) for i, t in enumerate(scheduler.timesteps ): _UpperCAmelCase : Optional[int] = model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample assert sample.dtype == torch.floataa
369
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : int = "speech_to_text_2" lowerCAmelCase : str = ["past_key_values"] lowerCAmelCase : int = {"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[Any] , lowerCamelCase__ : Tuple=1_00_00 , lowerCamelCase__ : Any=6 , lowerCamelCase__ : Tuple=20_48 , lowerCamelCase__ : List[Any]=4 , lowerCamelCase__ : Tuple=0.0 , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Tuple="relu" , lowerCamelCase__ : Dict=2_56 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : List[Any]=0.0 , lowerCamelCase__ : Optional[int]=0.0 , lowerCamelCase__ : List[Any]=0.0_2 , lowerCamelCase__ : Tuple=2 , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Any=1 , lowerCamelCase__ : int=0 , lowerCamelCase__ : str=2 , lowerCamelCase__ : List[Any]=10_24 , **lowerCamelCase__ : str , ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = vocab_size _UpperCAmelCase : Optional[int] = d_model _UpperCAmelCase : List[Any] = decoder_ffn_dim _UpperCAmelCase : Any = decoder_layers _UpperCAmelCase : int = decoder_attention_heads _UpperCAmelCase : Any = dropout _UpperCAmelCase : List[Any] = attention_dropout _UpperCAmelCase : Optional[int] = activation_dropout _UpperCAmelCase : List[Any] = activation_function _UpperCAmelCase : int = init_std _UpperCAmelCase : Dict = decoder_layerdrop _UpperCAmelCase : str = use_cache _UpperCAmelCase : Union[str, Any] = decoder_layers _UpperCAmelCase : Optional[Any] = scale_embedding # scale factor will be sqrt(d_model) if True _UpperCAmelCase : Any = max_target_positions super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
322
0
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : List[Any] = ["image_processor", "tokenizer"] lowerCAmelCase : Optional[Any] = "ChineseCLIPImageProcessor" lowerCAmelCase : Tuple = ("BertTokenizer", "BertTokenizerFast") def __init__( self : Tuple , lowerCamelCase__ : Any=None , lowerCamelCase__ : str=None , **lowerCamelCase__ : Tuple ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[str] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowerCamelCase__ , ) _UpperCAmelCase : List[Any] = kwargs.pop("feature_extractor" ) _UpperCAmelCase : List[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : int = self.image_processor def __call__( self : List[str] , lowerCamelCase__ : Optional[Any]=None , lowerCamelCase__ : Optional[int]=None , lowerCamelCase__ : Optional[int]=None , **lowerCamelCase__ : int ) ->int: '''simple docstring''' if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: _UpperCAmelCase : Any = self.tokenizer(lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ ) if images is not None: _UpperCAmelCase : Optional[Any] = self.image_processor(lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ ) if text is not None and images is not None: _UpperCAmelCase : Any = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowerCamelCase__ ) , tensor_type=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[Any] , *lowerCamelCase__ : Dict , **lowerCamelCase__ : Dict ) ->Optional[Any]: '''simple docstring''' return self.tokenizer.batch_decode(*lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] , *lowerCamelCase__ : List[Any] , **lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' return self.tokenizer.decode(*lowerCamelCase__ , **lowerCamelCase__ ) @property def lowerCAmelCase__ ( self : str ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Any = self.tokenizer.model_input_names _UpperCAmelCase : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowerCAmelCase__ ( self : int ) ->Union[str, Any]: '''simple docstring''' warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , lowerCamelCase__ , ) return self.image_processor_class
370
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = 'cuda' if torch.cuda.is_available() else 'cpu' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=100 , __lowerCAmelCase=" " ): _UpperCAmelCase : Any = text.split(__lowerCAmelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(__lowerCAmelCase ) , __lowerCAmelCase )] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase , _UpperCAmelCase : Dict = [], [] for title, text in zip(documents["title"] , documents["text"] ): if text is not None: for passage in split_text(__lowerCAmelCase ): titles.append(title if title is not None else "" ) texts.append(__lowerCAmelCase ) return {"title": titles, "text": texts} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : str = ctx_tokenizer( documents["title"] , documents["text"] , truncation=__lowerCAmelCase , padding="longest" , return_tensors="pt" )["input_ids"] _UpperCAmelCase : str = ctx_encoder(input_ids.to(device=__lowerCAmelCase ) , return_dict=__lowerCAmelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): ###################################### logger.info("Step 1 - Create the dataset" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way _UpperCAmelCase : Optional[int] = load_dataset( "csv" , data_files=[rag_example_args.csv_path] , split="train" , delimiter="\t" , column_names=["title", "text"] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words _UpperCAmelCase : Optional[int] = dataset.map(__lowerCAmelCase , batched=__lowerCAmelCase , num_proc=processing_args.num_proc ) # And compute the embeddings _UpperCAmelCase : Union[str, Any] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) _UpperCAmelCase : Dict = Features( {"text": Value("string" ), "title": Value("string" ), "embeddings": Sequence(Value("float32" ) )} ) # optional, save as float32 instead of float64 to save space _UpperCAmelCase : int = dataset.map( partial(__lowerCAmelCase , ctx_encoder=__lowerCAmelCase , ctx_tokenizer=__lowerCAmelCase ) , batched=__lowerCAmelCase , batch_size=processing_args.batch_size , features=__lowerCAmelCase , ) # And finally save your dataset _UpperCAmelCase : List[Any] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset" ) dataset.save_to_disk(__lowerCAmelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("Step 2 - Index the dataset" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search _UpperCAmelCase : Any = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("embeddings" , custom_index=__lowerCAmelCase ) # And save the index _UpperCAmelCase : List[str] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset_hnsw_index.faiss" ) dataset.get_index("embeddings" ).save(__lowerCAmelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" / "my_knowledge_dataset.csv" ) , metadata={"help": "Path to a tab-separated csv file with columns 'title' and 'text'"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Question that is passed as input to RAG. Default is 'What does Moses' rod turn into ?'."} , ) lowerCAmelCase : str = field( default="facebook/rag-sequence-nq" , metadata={"help": "The RAG model to use. Either 'facebook/rag-sequence-nq' or 'facebook/rag-token-nq'"} , ) lowerCAmelCase : str = field( default="facebook/dpr-ctx_encoder-multiset-base" , metadata={ "help": ( "The DPR context encoder model to use. Either 'facebook/dpr-ctx_encoder-single-nq-base' or" " 'facebook/dpr-ctx_encoder-multiset-base'" ) } , ) lowerCAmelCase : Optional[str] = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" ) , metadata={"help": "Path to a directory where the dataset passages and the index will be saved"} , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": "The number of processes to use to split the documents into passages. Default is single process." } , ) lowerCAmelCase : int = field( default=16 , metadata={ "help": "The batch size to use when computing the passages embeddings using the DPR context encoder." } , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : int = field( default=768 , metadata={"help": "The dimension of the embeddings to pass to the HNSW Faiss index."} , ) lowerCAmelCase : int = field( default=128 , metadata={ "help": ( "The number of bi-directional links created for every new element during the HNSW index construction." ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
322
0
import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser lowerCamelCase__ = logging.getLogger(__name__) torch.set_grad_enabled(False) lowerCamelCase__ = 'cuda' if torch.cuda.is_available() else 'cpu' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase=100 , __lowerCAmelCase=" " ): _UpperCAmelCase : Any = text.split(__lowerCAmelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(__lowerCAmelCase ) , __lowerCAmelCase )] def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : Dict = [], [] for title, text in zip(documents["title"] , documents["text"] ): if text is not None: for passage in split_text(__lowerCAmelCase ): titles.append(title if title is not None else "" ) texts.append(__lowerCAmelCase ) return {"title": titles, "text": texts} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : str = ctx_tokenizer( documents["title"] , documents["text"] , truncation=__lowerCAmelCase , padding="longest" , return_tensors="pt" )["input_ids"] _UpperCAmelCase : str = ctx_encoder(input_ids.to(device=__lowerCAmelCase ) , return_dict=__lowerCAmelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): ###################################### logger.info("Step 1 - Create the dataset" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way _UpperCAmelCase : Optional[int] = load_dataset( "csv" , data_files=[rag_example_args.csv_path] , split="train" , delimiter="\t" , column_names=["title", "text"] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words _UpperCAmelCase : Optional[int] = dataset.map(__lowerCAmelCase , batched=__lowerCAmelCase , num_proc=processing_args.num_proc ) # And compute the embeddings _UpperCAmelCase : Union[str, Any] = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=__lowerCAmelCase ) _UpperCAmelCase : Optional[int] = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) _UpperCAmelCase : Dict = Features( {"text": Value("string" ), "title": Value("string" ), "embeddings": Sequence(Value("float32" ) )} ) # optional, save as float32 instead of float64 to save space _UpperCAmelCase : int = dataset.map( partial(__lowerCAmelCase , ctx_encoder=__lowerCAmelCase , ctx_tokenizer=__lowerCAmelCase ) , batched=__lowerCAmelCase , batch_size=processing_args.batch_size , features=__lowerCAmelCase , ) # And finally save your dataset _UpperCAmelCase : List[Any] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset" ) dataset.save_to_disk(__lowerCAmelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("Step 2 - Index the dataset" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search _UpperCAmelCase : Any = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("embeddings" , custom_index=__lowerCAmelCase ) # And save the index _UpperCAmelCase : List[str] = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset_hnsw_index.faiss" ) dataset.get_index("embeddings" ).save(__lowerCAmelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" / "my_knowledge_dataset.csv" ) , metadata={"help": "Path to a tab-separated csv file with columns 'title' and 'text'"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Question that is passed as input to RAG. Default is 'What does Moses' rod turn into ?'."} , ) lowerCAmelCase : str = field( default="facebook/rag-sequence-nq" , metadata={"help": "The RAG model to use. Either 'facebook/rag-sequence-nq' or 'facebook/rag-token-nq'"} , ) lowerCAmelCase : str = field( default="facebook/dpr-ctx_encoder-multiset-base" , metadata={ "help": ( "The DPR context encoder model to use. Either 'facebook/dpr-ctx_encoder-single-nq-base' or" " 'facebook/dpr-ctx_encoder-multiset-base'" ) } , ) lowerCAmelCase : Optional[str] = field( default=str(Path(UpperCAmelCase__ ).parent / "test_run" / "dummy-kb" ) , metadata={"help": "Path to a directory where the dataset passages and the index will be saved"} , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": "The number of processes to use to split the documents into passages. Default is single process." } , ) lowerCAmelCase : int = field( default=16 , metadata={ "help": "The batch size to use when computing the passages embeddings using the DPR context encoder." } , ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : int = field( default=768 , metadata={"help": "The dimension of the embeddings to pass to the HNSW Faiss index."} , ) lowerCAmelCase : int = field( default=128 , metadata={ "help": ( "The number of bi-directional links created for every new element during the HNSW index construction." ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) lowerCamelCase__ = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: lowerCamelCase__ = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
371
'''simple docstring''' import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCamelCase__ = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class lowerCAmelCase__ ( unittest.TestCase ): @classmethod def lowerCAmelCase__ ( cls : List[str] ) ->str: '''simple docstring''' _UpperCAmelCase : Tuple = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def lowerCAmelCase__ ( cls : Union[str, Any] ) ->int: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def lowerCAmelCase__ ( self : int ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase__ , repo_id="test-config" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : Dict = BertConfig.from_pretrained(F"""{USER}/test-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) _UpperCAmelCase : List[str] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase__ , repo_id="valid_org/test-config-org" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) _UpperCAmelCase : int = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase__ , getattr(lowerCamelCase__ , lowerCamelCase__ ) ) def lowerCAmelCase__ ( self : List[str] ) ->Any: '''simple docstring''' CustomConfig.register_for_auto_class() _UpperCAmelCase : int = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) _UpperCAmelCase : str = AutoConfig.from_pretrained(F"""{USER}/test-dynamic-config""" , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class lowerCAmelCase__ ( unittest.TestCase ): def lowerCAmelCase__ ( self : List[str] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated _UpperCAmelCase : Any = c.n_embd + 1 # int _UpperCAmelCase : List[Any] = c.resid_pdrop + 1.0 # float _UpperCAmelCase : Tuple = not c.scale_attn_weights # bool _UpperCAmelCase : List[Any] = c.summary_type + "foo" # str c.update_from_string( F"""n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}""" ) self.assertEqual(lowerCamelCase__ , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(lowerCamelCase__ , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(lowerCamelCase__ , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(lowerCamelCase__ , c.summary_type , "mismatch for key: summary_type" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = PretrainedConfig() _UpperCAmelCase : Tuple = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase__ , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) _UpperCAmelCase : List[str] = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase__ , lowerCamelCase__ )] if len(lowerCamelCase__ ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" F""" {', '.join(lowerCamelCase__ )}.""" ) def lowerCAmelCase__ ( self : Optional[int] ) ->int: '''simple docstring''' with self.assertRaises(lowerCamelCase__ ): # config is in subfolder, the following should not work without specifying the subfolder _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) _UpperCAmelCase : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = mock.Mock() _UpperCAmelCase : List[str] = 5_00 _UpperCAmelCase : Dict = {} _UpperCAmelCase : Tuple = HTTPError _UpperCAmelCase : Any = {} # Download this model to make sure it's in the cache. _UpperCAmelCase : int = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=lowerCamelCase__ ) as mock_head: _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def lowerCAmelCase__ ( self : Optional[int] ) ->Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = AutoConfig.from_pretrained("bert-base-cased" ) _UpperCAmelCase : str = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase__ ) _UpperCAmelCase : Dict = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 _UpperCAmelCase : Optional[int] = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 _UpperCAmelCase : Dict = ["config.42.0.0.json"] _UpperCAmelCase : Union[str, Any] = 7_68 configuration.save_pretrained(lowerCamelCase__ ) shutil.move(os.path.join(lowerCamelCase__ , "config.4.0.0.json" ) , os.path.join(lowerCamelCase__ , "config.42.0.0.json" ) ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = "hf-internal-testing/test-two-configs" import transformers as new_transformers _UpperCAmelCase : Any = "v4.0.0" _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase__ , return_unused_kwargs=lowerCamelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers _UpperCAmelCase : List[Any] = "v3.0.0" _UpperCAmelCase : int = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertEqual(old_configuration.hidden_size , 7_68 )
322
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { 'tiiuae/falcon-40b': 'https://huggingface.co/tiiuae/falcon-40b/resolve/main/config.json', 'tiiuae/falcon-7b': 'https://huggingface.co/tiiuae/falcon-7b/resolve/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Dict = "falcon" lowerCAmelCase : Union[str, Any] = ["past_key_values"] def __init__( self : Optional[Any] , lowerCamelCase__ : Any=6_50_24 , lowerCamelCase__ : Optional[Any]=45_44 , lowerCamelCase__ : Union[str, Any]=32 , lowerCamelCase__ : List[Any]=71 , lowerCamelCase__ : Any=1E-5 , lowerCamelCase__ : Tuple=0.0_2 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : List[str]=0.0 , lowerCamelCase__ : Union[str, Any]=0.0 , lowerCamelCase__ : Optional[int]=None , lowerCamelCase__ : Dict=False , lowerCamelCase__ : Dict=False , lowerCamelCase__ : Union[str, Any]=True , lowerCamelCase__ : int=True , lowerCamelCase__ : List[Any]=False , lowerCamelCase__ : str=11 , lowerCamelCase__ : List[Any]=11 , **lowerCamelCase__ : List[str] , ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[Any] = vocab_size # Backward compatibility with n_embed kwarg _UpperCAmelCase : Any = kwargs.pop("n_embed" , lowerCamelCase__ ) _UpperCAmelCase : Dict = hidden_size if n_embed is None else n_embed _UpperCAmelCase : List[Any] = num_hidden_layers _UpperCAmelCase : Union[str, Any] = num_attention_heads _UpperCAmelCase : Tuple = layer_norm_epsilon _UpperCAmelCase : Any = initializer_range _UpperCAmelCase : int = use_cache _UpperCAmelCase : List[str] = hidden_dropout _UpperCAmelCase : int = attention_dropout _UpperCAmelCase : List[str] = bos_token_id _UpperCAmelCase : Tuple = eos_token_id _UpperCAmelCase : List[Any] = num_attention_heads if num_kv_heads is None else num_kv_heads _UpperCAmelCase : Union[str, Any] = alibi _UpperCAmelCase : str = new_decoder_architecture _UpperCAmelCase : int = multi_query # Ignored when new_decoder_architecture is True _UpperCAmelCase : Any = parallel_attn _UpperCAmelCase : Tuple = bias super().__init__(bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ ) @property def lowerCAmelCase__ ( self : Any ) ->str: '''simple docstring''' return self.hidden_size // self.num_attention_heads @property def lowerCAmelCase__ ( self : str ) ->str: '''simple docstring''' return not self.alibi
350
'''simple docstring''' from manim import * class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : List[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = Rectangle(height=0.5 , width=0.5 ) _UpperCAmelCase : Optional[Any] = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0 ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Dict = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[Any] = VGroup(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("CPU" , font_size=24 ) _UpperCAmelCase : Any = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(1 )] _UpperCAmelCase : str = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : int = Text("GPU" , font_size=24 ) _UpperCAmelCase : str = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) gpu.align_to(lowerCamelCase__ , lowerCamelCase__ ) gpu.set_x(gpu.get_x() - 1 ) self.add(lowerCamelCase__ ) _UpperCAmelCase : List[str] = [mem.copy() for i in range(6 )] _UpperCAmelCase : Any = VGroup(*lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0 ) _UpperCAmelCase : Optional[int] = Text("Model" , font_size=24 ) _UpperCAmelCase : Tuple = Group(lowerCamelCase__ , lowerCamelCase__ ).arrange(lowerCamelCase__ , buff=0.5 , aligned_edge=lowerCamelCase__ ) model.move_to([3, -1.0, 0] ) self.play( Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , Create(lowerCamelCase__ , run_time=1 ) , ) _UpperCAmelCase : int = MarkupText( F"""First, an empty model skeleton is loaded\ninto <span fgcolor='{YELLOW}'>memory</span> without using much RAM.""" , font_size=24 , ) _UpperCAmelCase : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _UpperCAmelCase : Union[str, Any] = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase__ , run_time=2.5 ) , Write(lowerCamelCase__ ) , Write(lowerCamelCase__ ) ) self.add(lowerCamelCase__ ) _UpperCAmelCase : int = [] _UpperCAmelCase : List[str] = [] _UpperCAmelCase : Dict = [] for i, rect in enumerate(lowerCamelCase__ ): _UpperCAmelCase : int = Rectangle(height=0.4_6 , width=0.4_6 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase__ , opacity=0.7 ) cpu_target.move_to(lowerCamelCase__ ) cpu_target.generate_target() _UpperCAmelCase : Dict = 0.4_6 / 4 _UpperCAmelCase : Any = 0.4_6 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.0_2 , direction=lowerCamelCase__ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=lowerCamelCase__ , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=lowerCamelCase__ , buff=0.0 ) cpu_targs.append(lowerCamelCase__ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(lowerCamelCase__ ) ) second_animations.append(MoveToTarget(lowerCamelCase__ , run_time=1.5 ) ) self.play(*lowerCamelCase__ ) self.play(*lowerCamelCase__ ) self.wait()
322
0
'''simple docstring''' import math def __lowerCAmelCase (): _UpperCAmelCase : str = input("Enter message: " ) _UpperCAmelCase : Union[str, Any] = int(input(F"""Enter key [2-{len(__lowerCAmelCase ) - 1}]: """ ) ) _UpperCAmelCase : List[str] = input("Encryption/Decryption [e/d]: " ) if mode.lower().startswith("e" ): _UpperCAmelCase : int = encrypt_message(__lowerCAmelCase , __lowerCAmelCase ) elif mode.lower().startswith("d" ): _UpperCAmelCase : Optional[int] = decrypt_message(__lowerCAmelCase , __lowerCAmelCase ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(F"""Output:\n{text + '|'}""" ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = [""] * key for col in range(__lowerCAmelCase ): _UpperCAmelCase : Dict = col while pointer < len(__lowerCAmelCase ): cipher_text[col] += message[pointer] pointer += key return "".join(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = math.ceil(len(__lowerCAmelCase ) / key ) _UpperCAmelCase : str = key _UpperCAmelCase : List[Any] = (num_cols * num_rows) - len(__lowerCAmelCase ) _UpperCAmelCase : str = [""] * num_cols _UpperCAmelCase : List[Any] = 0 _UpperCAmelCase : List[Any] = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): _UpperCAmelCase : int = 0 row += 1 return "".join(__lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
351
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1_024 , __lowerCAmelCase=1_024 , __lowerCAmelCase=False , **__lowerCAmelCase ): _UpperCAmelCase : Any = AutoTokenizer.from_pretrained(__lowerCAmelCase ) _UpperCAmelCase : List[str] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="train" , **__lowerCAmelCase ) _UpperCAmelCase : Dict = tok.pad_token_id def get_lens(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = tqdm( DataLoader(__lowerCAmelCase , batch_size=512 , num_workers=8 , shuffle=__lowerCAmelCase , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) _UpperCAmelCase : List[str] = [] for batch in dl: _UpperCAmelCase : Any = batch["input_ids"].ne(__lowerCAmelCase ).sum(1 ).tolist() _UpperCAmelCase : Tuple = batch["labels"].ne(__lowerCAmelCase ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__lowerCAmelCase , __lowerCAmelCase ): max_lens.append(max(__lowerCAmelCase , __lowerCAmelCase ) ) else: max_lens.extend(__lowerCAmelCase ) return max_lens _UpperCAmelCase : Dict = get_lens(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = SeqaSeqDataset(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , type_path="val" , **__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = get_lens(__lowerCAmelCase ) pickle_save(__lowerCAmelCase , train_ds.len_file ) pickle_save(__lowerCAmelCase , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
322
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { 'configuration_jukebox': [ 'JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'JukeboxConfig', 'JukeboxPriorConfig', 'JukeboxVQVAEConfig', ], 'tokenization_jukebox': ['JukeboxTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ 'JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'JukeboxModel', 'JukeboxPreTrainedModel', 'JukeboxVQVAE', 'JukeboxPrior', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
352
'''simple docstring''' import pytest lowerCamelCase__ = '__dummy_dataset1__' lowerCamelCase__ = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def __lowerCAmelCase (): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = dataset_loading_script_name _UpperCAmelCase : Any = tmp_path / "datasets" / script_name script_dir.mkdir(parents=__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = script_dir / F"""{script_name}.py""" with open(__lowerCAmelCase , "w" ) as f: f.write(__lowerCAmelCase ) return str(__lowerCAmelCase )
322
0
'''simple docstring''' from __future__ import annotations import math def __lowerCAmelCase (__lowerCAmelCase ) -> Optional[int]: if num <= 0: _UpperCAmelCase : Any = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(__lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = [True] * (num + 1) _UpperCAmelCase : Any = [] _UpperCAmelCase : str = 2 _UpperCAmelCase : int = int(math.sqrt(__lowerCAmelCase ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(__lowerCAmelCase ) # Set multiples of start be False for i in range(start * start , num + 1 , __lowerCAmelCase ): if sieve[i] is True: _UpperCAmelCase : Optional[int] = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(__lowerCAmelCase ) return prime if __name__ == "__main__": print(prime_sieve(int(input('Enter a positive integer: ').strip())))
353
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase__ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCamelCase__ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCamelCase__ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCamelCase__ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py"] , reference_urls=[ "https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score", "https://en.wikipedia.org/wiki/METEOR", ] , ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : List[str] ) ->int: '''simple docstring''' import nltk nltk.download("wordnet" ) if NLTK_VERSION >= version.Version("3.6.5" ): nltk.download("punkt" ) if NLTK_VERSION >= version.Version("3.6.6" ): nltk.download("omw-1.4" ) def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : int=0.9 , lowerCamelCase__ : Dict=3 , lowerCamelCase__ : Dict=0.5 ) ->Any: '''simple docstring''' if NLTK_VERSION >= version.Version("3.6.5" ): _UpperCAmelCase : Dict = [ meteor_score.single_meteor_score( word_tokenize(lowerCamelCase__ ) , word_tokenize(lowerCamelCase__ ) , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] else: _UpperCAmelCase : Optional[int] = [ meteor_score.single_meteor_score(lowerCamelCase__ , lowerCamelCase__ , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , gamma=lowerCamelCase__ ) for ref, pred in zip(lowerCamelCase__ , lowerCamelCase__ ) ] return {"meteor": np.mean(lowerCamelCase__ )}
322
0
'''simple docstring''' import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1_024 ): _UpperCAmelCase : int = [], [] _UpperCAmelCase : Union[str, Any] = list(zip(__lowerCAmelCase , __lowerCAmelCase ) ) _UpperCAmelCase : List[Any] = sorted_examples[0] def is_too_big(__lowerCAmelCase ): return tok(__lowerCAmelCase , return_tensors="pt" ).input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:] ): _UpperCAmelCase : Optional[Any] = new_src + " " + src _UpperCAmelCase : Optional[int] = new_tgt + " " + tgt if is_too_big(__lowerCAmelCase ) or is_too_big(__lowerCAmelCase ): # cant fit, finalize example finished_src.append(__lowerCAmelCase ) finished_tgt.append(__lowerCAmelCase ) _UpperCAmelCase : str = src, tgt else: # can fit, keep adding _UpperCAmelCase : Any = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(__lowerCAmelCase ) finished_tgt.append(__lowerCAmelCase ) return finished_src, finished_tgt def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[int] = Path(__lowerCAmelCase ) save_path.mkdir(exist_ok=__lowerCAmelCase ) for split in ["train"]: _UpperCAmelCase : Optional[Any] = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" _UpperCAmelCase : List[Any] = [x.rstrip() for x in Path(__lowerCAmelCase ).open().readlines()] _UpperCAmelCase : Optional[int] = [x.rstrip() for x in Path(__lowerCAmelCase ).open().readlines()] _UpperCAmelCase : Union[str, Any] = pack_examples(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) print(F"""packed {split} split from {len(__lowerCAmelCase )} examples -> {len(__lowerCAmelCase )}.""" ) Path(save_path / F"""{split}.source""" ).open("w" ).write("\n".join(__lowerCAmelCase ) ) Path(save_path / F"""{split}.target""" ).open("w" ).write("\n".join(__lowerCAmelCase ) ) for split in ["val", "test"]: _UpperCAmelCase : Dict = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" shutil.copyfile(__lowerCAmelCase , save_path / F"""{split}.source""" ) shutil.copyfile(__lowerCAmelCase , save_path / F"""{split}.target""" ) def __lowerCAmelCase (): _UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument("--tok_name" , type=__lowerCAmelCase , help="like facebook/bart-large-cnn,t5-base, etc." ) parser.add_argument("--max_seq_len" , type=__lowerCAmelCase , default=128 ) parser.add_argument("--data_dir" , type=__lowerCAmelCase ) parser.add_argument("--save_path" , type=__lowerCAmelCase ) _UpperCAmelCase : List[Any] = parser.parse_args() _UpperCAmelCase : int = AutoTokenizer.from_pretrained(args.tok_name ) return pack_data_dir(__lowerCAmelCase , Path(args.data_dir ) , args.max_seq_len , args.save_path ) if __name__ == "__main__": packer_cli()
354
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline lowerCamelCase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : int ) ->str: '''simple docstring''' if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Union[str, Any] = [label.strip() for label in labels.split("," ) if label.strip()] return labels def __call__( self : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Any , lowerCamelCase__ : List[Any] ) ->str: '''simple docstring''' if len(lowerCamelCase__ ) == 0 or len(lowerCamelCase__ ) == 0: raise ValueError("You must include at least one label and at least one sequence." ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( "The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. " "Make sure the passed template includes formatting syntax such as {{}} where the label should go." ).format(lowerCamelCase__ ) ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _UpperCAmelCase : Optional[Any] = [sequences] _UpperCAmelCase : int = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(lowerCamelCase__ )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(UpperCAmelCase__ ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Union[str, Any] , lowerCamelCase__ : Optional[Any]=ZeroShotClassificationArgumentHandler() , *lowerCamelCase__ : List[str] , **lowerCamelCase__ : Any ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : Tuple = args_parser super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) if self.entailment_id == -1: logger.warning( "Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to " "-1. Define a descriptive label2id mapping in the model config to ensure correct outputs." ) @property def lowerCAmelCase__ ( self : Any ) ->Union[str, Any]: '''simple docstring''' for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("entail" ): return ind return -1 def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Tuple , lowerCamelCase__ : int=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : str=TruncationStrategy.ONLY_FIRST , **lowerCamelCase__ : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : int = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( "Tokenizer was not supporting padding necessary for zero-shot, attempting to use " " `pad_token=eos_token`" ) _UpperCAmelCase : Optional[Any] = self.tokenizer.eos_token try: _UpperCAmelCase : List[str] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , ) except Exception as e: if "too short" in str(lowerCamelCase__ ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. _UpperCAmelCase : List[Any] = self.tokenizer( lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def lowerCAmelCase__ ( self : int , **lowerCamelCase__ : Union[str, Any] ) ->Tuple: '''simple docstring''' if kwargs.get("multi_class" , lowerCamelCase__ ) is not None: _UpperCAmelCase : int = kwargs["multi_class"] logger.warning( "The `multi_class` argument has been deprecated and renamed to `multi_label`. " "`multi_class` will be removed in a future version of Transformers." ) _UpperCAmelCase : Dict = {} if "candidate_labels" in kwargs: _UpperCAmelCase : List[Any] = self._args_parser._parse_labels(kwargs["candidate_labels"] ) if "hypothesis_template" in kwargs: _UpperCAmelCase : Dict = kwargs["hypothesis_template"] _UpperCAmelCase : List[str] = {} if "multi_label" in kwargs: _UpperCAmelCase : Optional[Any] = kwargs["multi_label"] return preprocess_params, {}, postprocess_params def __call__( self : int , lowerCamelCase__ : Union[str, List[str]] , *lowerCamelCase__ : str , **lowerCamelCase__ : Optional[Any] , ) ->Optional[int]: '''simple docstring''' if len(lowerCamelCase__ ) == 0: pass elif len(lowerCamelCase__ ) == 1 and "candidate_labels" not in kwargs: _UpperCAmelCase : int = args[0] else: raise ValueError(F"""Unable to understand extra arguments {args}""" ) return super().__call__(lowerCamelCase__ , **lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Any=None , lowerCamelCase__ : str="This example is {}." ) ->Tuple: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Optional[int] = self._args_parser(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for i, (candidate_label, sequence_pair) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): _UpperCAmelCase : Optional[int] = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(lowerCamelCase__ ) - 1, **model_input, } def lowerCAmelCase__ ( self : Optional[Any] , lowerCamelCase__ : Optional[int] ) ->int: '''simple docstring''' _UpperCAmelCase : Dict = inputs["candidate_label"] _UpperCAmelCase : Optional[int] = inputs["sequence"] _UpperCAmelCase : Dict = {k: inputs[k] for k in self.tokenizer.model_input_names} _UpperCAmelCase : List[Any] = self.model(**lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = { "candidate_label": candidate_label, "sequence": sequence, "is_last": inputs["is_last"], **outputs, } return model_outputs def lowerCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : int , lowerCamelCase__ : Tuple=False ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Any = [outputs["candidate_label"] for outputs in model_outputs] _UpperCAmelCase : Any = [outputs["sequence"] for outputs in model_outputs] _UpperCAmelCase : Optional[int] = np.concatenate([output["logits"].numpy() for output in model_outputs] ) _UpperCAmelCase : Optional[Any] = logits.shape[0] _UpperCAmelCase : Any = len(lowerCamelCase__ ) _UpperCAmelCase : str = N // n _UpperCAmelCase : str = logits.reshape((num_sequences, n, -1) ) if multi_label or len(lowerCamelCase__ ) == 1: # softmax over the entailment vs. contradiction dim for each label independently _UpperCAmelCase : int = self.entailment_id _UpperCAmelCase : List[Any] = -1 if entailment_id == 0 else 0 _UpperCAmelCase : str = reshaped_outputs[..., [contradiction_id, entailment_id]] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : str = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels _UpperCAmelCase : int = reshaped_outputs[..., self.entailment_id] _UpperCAmelCase : Union[str, Any] = np.exp(lowerCamelCase__ ) / np.exp(lowerCamelCase__ ).sum(-1 , keepdims=lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
322
0
'''simple docstring''' import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = (KDPMaDiscreteScheduler,) lowerCAmelCase : Optional[int] = 10 def lowerCAmelCase__ ( self : Optional[int] , **lowerCamelCase__ : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : Optional[Any] = { "num_train_timesteps": 11_00, "beta_start": 0.0_0_0_1, "beta_end": 0.0_2, "beta_schedule": "linear", } config.update(**lowerCamelCase__ ) return config def lowerCAmelCase__ ( self : Union[str, Any] ) ->Optional[int]: '''simple docstring''' for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict ) ->str: '''simple docstring''' for beta_start, beta_end in zip([0.0_0_0_0_1, 0.0_0_0_1, 0.0_0_1] , [0.0_0_0_2, 0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=lowerCamelCase__ , beta_end=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Dict ) ->Tuple: '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase__ ) def lowerCAmelCase__ ( self : Union[str, Any] ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Dict = self.scheduler_classes[0] _UpperCAmelCase : Optional[int] = self.get_scheduler_config(prediction_type="v_prediction" ) _UpperCAmelCase : Tuple = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) _UpperCAmelCase : Dict = self.dummy_model() _UpperCAmelCase : str = self.dummy_sample_deter * scheduler.init_noise_sigma _UpperCAmelCase : List[Any] = sample.to(lowerCamelCase__ ) for i, t in enumerate(scheduler.timesteps ): _UpperCAmelCase : Optional[int] = scheduler.scale_model_input(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Dict = model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Tuple = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Any = output.prev_sample _UpperCAmelCase : Optional[Any] = torch.sum(torch.abs(lowerCamelCase__ ) ) _UpperCAmelCase : Tuple = torch.mean(torch.abs(lowerCamelCase__ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693428650170972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.0_0_0_2 ) < 1E-3 def lowerCAmelCase__ ( self : List[Any] ) ->Optional[Any]: '''simple docstring''' if torch_device == "mps": return _UpperCAmelCase : Union[str, Any] = self.scheduler_classes[0] _UpperCAmelCase : Any = self.get_scheduler_config() _UpperCAmelCase : Any = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) _UpperCAmelCase : Optional[Any] = self.dummy_model() _UpperCAmelCase : str = self.dummy_sample_deter * scheduler.init_noise_sigma _UpperCAmelCase : List[Any] = sample.to(lowerCamelCase__ ) for i, t in enumerate(scheduler.timesteps ): _UpperCAmelCase : List[Any] = scheduler.scale_model_input(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : str = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = output.prev_sample _UpperCAmelCase : Any = torch.sum(torch.abs(lowerCamelCase__ ) ) _UpperCAmelCase : List[str] = torch.mean(torch.abs(lowerCamelCase__ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 def lowerCAmelCase__ ( self : Any ) ->str: '''simple docstring''' if torch_device == "mps": return _UpperCAmelCase : Union[str, Any] = self.scheduler_classes[0] _UpperCAmelCase : Optional[int] = self.get_scheduler_config() _UpperCAmelCase : Any = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(self.num_inference_steps , device=lowerCamelCase__ ) _UpperCAmelCase : Tuple = self.dummy_model() _UpperCAmelCase : Dict = self.dummy_sample_deter.to(lowerCamelCase__ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: _UpperCAmelCase : Optional[int] = scheduler.scale_model_input(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Any = model(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[Any] = scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : Tuple = output.prev_sample _UpperCAmelCase : List[Any] = torch.sum(torch.abs(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[int] = torch.mean(torch.abs(lowerCamelCase__ ) ) if str(lowerCamelCase__ ).startswith("cpu" ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1E-3
355
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase = 4_000_000 ): _UpperCAmelCase : List[Any] = [] _UpperCAmelCase , _UpperCAmelCase : Dict = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__lowerCAmelCase ) _UpperCAmelCase , _UpperCAmelCase : Any = b, a + b return sum(__lowerCAmelCase ) if __name__ == "__main__": print(F'''{solution() = }''')
322
0
'''simple docstring''' import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : int , lowerCamelCase__ : str , lowerCamelCase__ : Optional[Any]=13 , lowerCamelCase__ : List[str]=30 , lowerCamelCase__ : str=2 , lowerCamelCase__ : Union[str, Any]=3 , lowerCamelCase__ : Tuple=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Tuple=32 , lowerCamelCase__ : Any=5 , lowerCamelCase__ : Tuple=4 , lowerCamelCase__ : Dict=37 , lowerCamelCase__ : str="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Optional[Any]=0.1 , lowerCamelCase__ : str=10 , lowerCamelCase__ : List[Any]=0.0_2 , ) ->Dict: '''simple docstring''' _UpperCAmelCase : Optional[int] = parent _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Tuple = image_size _UpperCAmelCase : str = patch_size _UpperCAmelCase : Any = num_channels _UpperCAmelCase : Optional[int] = is_training _UpperCAmelCase : Union[str, Any] = use_labels _UpperCAmelCase : List[str] = hidden_size _UpperCAmelCase : List[str] = num_hidden_layers _UpperCAmelCase : Dict = num_attention_heads _UpperCAmelCase : Any = intermediate_size _UpperCAmelCase : Any = hidden_act _UpperCAmelCase : List[str] = hidden_dropout_prob _UpperCAmelCase : Tuple = attention_probs_dropout_prob _UpperCAmelCase : Dict = type_sequence_label_size _UpperCAmelCase : Union[str, Any] = initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _UpperCAmelCase : Any = (image_size // patch_size) ** 2 _UpperCAmelCase : Optional[int] = num_patches + 1 def lowerCAmelCase__ ( self : List[Any] ) ->str: '''simple docstring''' _UpperCAmelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase : Any = ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, pixel_values def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : Tuple , lowerCamelCase__ : Tuple ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : Optional[int] = FlaxViTModel(config=lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = model(lowerCamelCase__ ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) _UpperCAmelCase : str = (self.image_size, self.image_size) _UpperCAmelCase : Any = (self.patch_size, self.patch_size) _UpperCAmelCase : str = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, num_patches + 1, self.hidden_size) ) def lowerCAmelCase__ ( self : Dict , lowerCamelCase__ : str , lowerCamelCase__ : str ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Any = self.type_sequence_label_size _UpperCAmelCase : Any = FlaxViTForImageClassification(config=lowerCamelCase__ ) _UpperCAmelCase : int = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _UpperCAmelCase : Optional[int] = 1 _UpperCAmelCase : str = FlaxViTForImageClassification(lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _UpperCAmelCase : Tuple = model(lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[str] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() ( _UpperCAmelCase ) : Union[str, Any] = config_and_inputs _UpperCAmelCase : int = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Union[str, Any] = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def lowerCAmelCase__ ( self : Any ) ->None: '''simple docstring''' _UpperCAmelCase : Any = FlaxViTModelTester(self ) _UpperCAmelCase : List[str] = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=37 ) def lowerCAmelCase__ ( self : Optional[int] ) ->List[Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCAmelCase__ ( self : str ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[Any] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) def lowerCAmelCase__ ( self : List[str] ) ->List[Any]: '''simple docstring''' _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : Dict = model_class(lowerCamelCase__ ) _UpperCAmelCase : Tuple = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase : Tuple = [*signature.parameters.keys()] _UpperCAmelCase : str = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def lowerCAmelCase__ ( self : int ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _UpperCAmelCase : Union[str, Any] = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _UpperCAmelCase : List[str] = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ : Optional[Any] , **lowerCamelCase__ : str ): return model(pixel_values=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest("JIT Enabled" ): _UpperCAmelCase : Union[str, Any] = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _UpperCAmelCase : List[Any] = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def lowerCAmelCase__ ( self : Union[str, Any] ) ->int: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCAmelCase : Optional[Any] = model_class_name.from_pretrained("google/vit-base-patch16-224" ) _UpperCAmelCase : Optional[Any] = model(np.ones((1, 3, 2_24, 2_24) ) ) self.assertIsNotNone(lowerCamelCase__ )
356
'''simple docstring''' import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class lowerCAmelCase__ ( unittest.TestCase ): def __init__( self : Optional[Any] , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : List[str]=13 , lowerCamelCase__ : Optional[Any]=7 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Any=True , lowerCamelCase__ : int=99 , lowerCamelCase__ : int=32 , lowerCamelCase__ : List[str]=5 , lowerCamelCase__ : Optional[Any]=4 , lowerCamelCase__ : Optional[int]=37 , lowerCamelCase__ : Tuple="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Union[str, Any]=0.1 , lowerCamelCase__ : Optional[int]=5_12 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : str=2 , lowerCamelCase__ : Union[str, Any]=0.0_2 , lowerCamelCase__ : Tuple=4 , ) ->Union[str, Any]: '''simple docstring''' _UpperCAmelCase : List[Any] = parent _UpperCAmelCase : List[Any] = batch_size _UpperCAmelCase : Optional[int] = seq_length _UpperCAmelCase : int = is_training _UpperCAmelCase : Dict = use_attention_mask _UpperCAmelCase : Optional[Any] = use_token_type_ids _UpperCAmelCase : int = use_labels _UpperCAmelCase : Optional[int] = vocab_size _UpperCAmelCase : Any = hidden_size _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : List[Any] = num_attention_heads _UpperCAmelCase : Tuple = intermediate_size _UpperCAmelCase : int = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob _UpperCAmelCase : Union[str, Any] = max_position_embeddings _UpperCAmelCase : Tuple = type_vocab_size _UpperCAmelCase : List[Any] = type_sequence_label_size _UpperCAmelCase : Optional[int] = initializer_range _UpperCAmelCase : Dict = num_choices def lowerCAmelCase__ ( self : List[Any] ) ->Any: '''simple docstring''' _UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Dict = None if self.use_attention_mask: _UpperCAmelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCAmelCase : Union[str, Any] = None if self.use_token_type_ids: _UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _UpperCAmelCase : int = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Tuple = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = config_and_inputs _UpperCAmelCase : str = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCAmelCase : Optional[int] = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase__ ( self : Optional[int] ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : int = FlaxAlbertModelTester(self ) @slow def lowerCAmelCase__ ( self : Any ) ->List[str]: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCAmelCase : List[str] = model_class_name.from_pretrained("albert-base-v2" ) _UpperCAmelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): @slow def lowerCAmelCase__ ( self : Tuple ) ->Optional[Any]: '''simple docstring''' _UpperCAmelCase : str = FlaxAlbertModel.from_pretrained("albert-base-v2" ) _UpperCAmelCase : List[Any] = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) _UpperCAmelCase : Optional[int] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCAmelCase : Dict = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )[0] _UpperCAmelCase : List[Any] = (1, 11, 7_68) self.assertEqual(output.shape , lowerCamelCase__ ) _UpperCAmelCase : str = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase__ , atol=1E-4 ) )
322
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ : int = logging.get_logger(__name__) lowerCamelCase__ : Optional[Any] = { 'microsoft/markuplm-base': 'https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json', 'microsoft/markuplm-large': 'https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json', } class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : List[Any] = "markuplm" def __init__( self : Optional[int] , lowerCamelCase__ : Optional[int]=3_05_22 , lowerCamelCase__ : int=7_68 , lowerCamelCase__ : List[Any]=12 , lowerCamelCase__ : Optional[int]=12 , lowerCamelCase__ : List[str]=30_72 , lowerCamelCase__ : Optional[Any]="gelu" , lowerCamelCase__ : Any=0.1 , lowerCamelCase__ : Tuple=0.1 , lowerCamelCase__ : str=5_12 , lowerCamelCase__ : Optional[int]=2 , lowerCamelCase__ : int=0.0_2 , lowerCamelCase__ : List[Any]=1E-12 , lowerCamelCase__ : str=0 , lowerCamelCase__ : List[str]=0 , lowerCamelCase__ : Optional[int]=2 , lowerCamelCase__ : List[Any]=2_56 , lowerCamelCase__ : List[Any]=10_24 , lowerCamelCase__ : Optional[Any]=2_16 , lowerCamelCase__ : Any=10_01 , lowerCamelCase__ : List[str]=32 , lowerCamelCase__ : Union[str, Any]=50 , lowerCamelCase__ : Optional[Any]="absolute" , lowerCamelCase__ : Any=True , lowerCamelCase__ : List[str]=None , **lowerCamelCase__ : Optional[Any] , ) ->Tuple: '''simple docstring''' super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ , ) _UpperCAmelCase : Union[str, Any] = vocab_size _UpperCAmelCase : List[str] = hidden_size _UpperCAmelCase : Tuple = num_hidden_layers _UpperCAmelCase : Tuple = num_attention_heads _UpperCAmelCase : List[Any] = hidden_act _UpperCAmelCase : List[str] = intermediate_size _UpperCAmelCase : Optional[Any] = hidden_dropout_prob _UpperCAmelCase : Any = attention_probs_dropout_prob _UpperCAmelCase : List[str] = max_position_embeddings _UpperCAmelCase : List[str] = type_vocab_size _UpperCAmelCase : Any = initializer_range _UpperCAmelCase : Union[str, Any] = layer_norm_eps _UpperCAmelCase : Optional[Any] = position_embedding_type _UpperCAmelCase : Any = use_cache _UpperCAmelCase : Tuple = classifier_dropout # additional properties _UpperCAmelCase : Dict = max_depth _UpperCAmelCase : Optional[int] = max_xpath_tag_unit_embeddings _UpperCAmelCase : List[Any] = max_xpath_subs_unit_embeddings _UpperCAmelCase : Any = tag_pad_id _UpperCAmelCase : Any = subs_pad_id _UpperCAmelCase : Dict = xpath_unit_hidden_size
357
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowerCamelCase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-classification/requirements.txt') lowerCamelCase__ = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) lowerCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def __lowerCAmelCase (__lowerCAmelCase ): with open(__lowerCAmelCase , "rb" ) as f: _UpperCAmelCase : List[str] = Image.open(__lowerCAmelCase ) return im.convert("RGB" ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={ "help": "Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub)." } , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the training data."} ) lowerCAmelCase : Optional[str] = field(default=UpperCAmelCase__ , metadata={"help": "A folder containing the validation data."} ) lowerCAmelCase : Optional[float] = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCAmelCase : Optional[int] = field( default=UpperCAmelCase__ , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def lowerCAmelCase__ ( self : int ) ->List[str]: '''simple docstring''' if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class lowerCAmelCase__ : lowerCAmelCase : str = field( default="google/vit-base-patch16-224-in21k" , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCAmelCase__ )} , ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCAmelCase : Optional[str] = field( default=UpperCAmelCase__ , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) lowerCAmelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCAmelCase : str = field(default=UpperCAmelCase__ , metadata={"help": "Name or path of preprocessor config."} ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) lowerCAmelCase : bool = field( default=UpperCAmelCase__ , metadata={"help": "Will enable to load a pretrained model whose head dimensions are different."} , ) def __lowerCAmelCase (__lowerCAmelCase ): _UpperCAmelCase : str = torch.stack([example["pixel_values"] for example in examples] ) _UpperCAmelCase : Tuple = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def __lowerCAmelCase (): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , __lowerCAmelCase , __lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase : Optional[Any] = training_args.get_process_log_level() logger.setLevel(__lowerCAmelCase ) transformers.utils.logging.set_verbosity(__lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _UpperCAmelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _UpperCAmelCase : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: _UpperCAmelCase : List[Any] = {} if data_args.train_dir is not None: _UpperCAmelCase : str = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: _UpperCAmelCase : Optional[Any] = os.path.join(data_args.validation_dir , "**" ) _UpperCAmelCase : Any = load_dataset( "imagefolder" , data_files=__lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. _UpperCAmelCase : int = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __lowerCAmelCase ) and data_args.train_val_split > 0.0: _UpperCAmelCase : List[Any] = dataset["train"].train_test_split(data_args.train_val_split ) _UpperCAmelCase : List[str] = split["train"] _UpperCAmelCase : Union[str, Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase : Optional[int] = dataset["train"].features["labels"].names _UpperCAmelCase , _UpperCAmelCase : int = {}, {} for i, label in enumerate(__lowerCAmelCase ): _UpperCAmelCase : int = str(__lowerCAmelCase ) _UpperCAmelCase : str = label # Load the accuracy metric from the datasets package _UpperCAmelCase : int = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(__lowerCAmelCase ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _UpperCAmelCase : Dict = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowerCAmelCase ) , labelaid=__lowerCAmelCase , idalabel=__lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _UpperCAmelCase : Dict = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _UpperCAmelCase : int = image_processor.size["shortest_edge"] else: _UpperCAmelCase : int = (image_processor.size["height"], image_processor.size["width"]) _UpperCAmelCase : str = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _UpperCAmelCase : Optional[int] = Compose( [ RandomResizedCrop(__lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _UpperCAmelCase : Union[str, Any] = Compose( [ Resize(__lowerCAmelCase ), CenterCrop(__lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(__lowerCAmelCase ): _UpperCAmelCase : Optional[Any] = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(__lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: _UpperCAmelCase : Dict = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(__lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: _UpperCAmelCase : Optional[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(__lowerCAmelCase ) # Initalize our trainer _UpperCAmelCase : Union[str, Any] = Trainer( model=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=__lowerCAmelCase , tokenizer=__lowerCAmelCase , data_collator=__lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : Any = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : int = last_checkpoint _UpperCAmelCase : Dict = trainer.train(resume_from_checkpoint=__lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase : Dict = trainer.evaluate() trainer.log_metrics("eval" , __lowerCAmelCase ) trainer.save_metrics("eval" , __lowerCAmelCase ) # Write model card and (optionally) push to hub _UpperCAmelCase : int = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowerCAmelCase ) else: trainer.create_model_card(**__lowerCAmelCase ) if __name__ == "__main__": main()
322
0
from queue import PriorityQueue from typing import Any import numpy as np def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ): for nxt, d in graph[v]: if nxt in visited_forward: continue _UpperCAmelCase : Dict = cst_fwd.get(__lowerCAmelCase , np.inf ) _UpperCAmelCase : List[Any] = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) _UpperCAmelCase : int = new_cost_f _UpperCAmelCase : str = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: _UpperCAmelCase : Tuple = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Optional[int] = -1 _UpperCAmelCase : Optional[int] = set() _UpperCAmelCase : List[Any] = set() _UpperCAmelCase : str = {source: 0} _UpperCAmelCase : Union[str, Any] = {destination: 0} _UpperCAmelCase : Union[str, Any] = {source: None} _UpperCAmelCase : List[Any] = {destination: None} _UpperCAmelCase : PriorityQueue[Any] = PriorityQueue() _UpperCAmelCase : PriorityQueue[Any] = PriorityQueue() _UpperCAmelCase : Union[str, Any] = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): _UpperCAmelCase : Dict = queue_forward.get() visited_forward.add(__lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = queue_backward.get() visited_backward.add(__lowerCAmelCase ) _UpperCAmelCase : List[str] = pass_and_relaxation( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ) _UpperCAmelCase : Optional[int] = pass_and_relaxation( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: _UpperCAmelCase : str = shortest_distance return shortest_path_distance lowerCamelCase__ = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } lowerCamelCase__ = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
358
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig lowerCamelCase__ = logging.get_logger(__name__) # General docstring lowerCamelCase__ = 'RegNetConfig' # Base docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = [1, 1_088, 7, 7] # Image classification docstring lowerCamelCase__ = 'facebook/regnet-y-040' lowerCamelCase__ = 'tabby, tabby cat' lowerCamelCase__ = [ 'facebook/regnet-y-040', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 3 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : int = 1 , lowerCamelCase__ : Optional[str] = "relu" , **lowerCamelCase__ : Tuple , ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _UpperCAmelCase : Optional[Any] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _UpperCAmelCase : Dict = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=lowerCamelCase__ , strides=lowerCamelCase__ , padding="VALID" , groups=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" , ) _UpperCAmelCase : List[Any] = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) _UpperCAmelCase : int = ACTaFN[activation] if activation is not None else tf.identity def lowerCAmelCase__ ( self : int , lowerCamelCase__ : Tuple ) ->Any: '''simple docstring''' _UpperCAmelCase : List[str] = self.convolution(self.padding(lowerCamelCase__ ) ) _UpperCAmelCase : Optional[Any] = self.normalization(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : Optional[Any] ) ->Optional[Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = config.num_channels _UpperCAmelCase : Any = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="embedder" , ) def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : Optional[Any] ) ->Dict: '''simple docstring''' _UpperCAmelCase : List[str] = shape_list(lowerCamelCase__ )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _UpperCAmelCase : Optional[Any] = tf.transpose(lowerCamelCase__ , perm=(0, 2, 3, 1) ) _UpperCAmelCase : List[Any] = self.embedder(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : int ) ->Union[str, Any]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : int = tf.keras.layers.ConvaD( filters=lowerCamelCase__ , kernel_size=1 , strides=lowerCamelCase__ , use_bias=lowerCamelCase__ , name="convolution" ) _UpperCAmelCase : Any = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="normalization" ) def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False ) ->tf.Tensor: '''simple docstring''' return self.normalization(self.convolution(lowerCamelCase__ ) , training=lowerCamelCase__ ) class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Any , lowerCamelCase__ : int , lowerCamelCase__ : int , **lowerCamelCase__ : Optional[int] ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) _UpperCAmelCase : int = [ tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="relu" , name="attention.0" ), tf.keras.layers.ConvaD(filters=lowerCamelCase__ , kernel_size=1 , activation="sigmoid" , name="attention.2" ), ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Optional[int] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.pooler(lowerCamelCase__ ) for layer_module in self.attention: _UpperCAmelCase : str = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = hidden_state * pooled return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : Any ) ->List[str]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : List[str] = in_channels != out_channels or stride != 1 _UpperCAmelCase : List[str] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : List[str] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _UpperCAmelCase : Optional[int] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.2" ), ] _UpperCAmelCase : Union[str, Any] = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Tuple , lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' _UpperCAmelCase : Any = hidden_state for layer_module in self.layers: _UpperCAmelCase : List[Any] = layer_module(lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : List[Any] = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : List[Any] , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 1 , **lowerCamelCase__ : str ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = in_channels != out_channels or stride != 1 _UpperCAmelCase : Optional[int] = max(1 , out_channels // config.groups_width ) _UpperCAmelCase : Union[str, Any] = ( TFRegNetShortCut(lowerCamelCase__ , stride=lowerCamelCase__ , name="shortcut" ) if should_apply_shortcut else tf.keras.layers.Activation("linear" , name="shortcut" ) ) _UpperCAmelCase : List[Any] = [ TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act , name="layer.0" ), TFRegNetConvLayer( lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act , name="layer.1" ), TFRegNetSELayer(lowerCamelCase__ , reduced_channels=int(round(in_channels / 4 ) ) , name="layer.2" ), TFRegNetConvLayer(lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ , name="layer.3" ), ] _UpperCAmelCase : int = ACTaFN[config.hidden_act] def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : str ) ->Any: '''simple docstring''' _UpperCAmelCase : int = hidden_state for layer_module in self.layers: _UpperCAmelCase : Tuple = layer_module(lowerCamelCase__ ) _UpperCAmelCase : List[Any] = self.shortcut(lowerCamelCase__ ) hidden_state += residual _UpperCAmelCase : Tuple = self.activation(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , lowerCamelCase__ : int , lowerCamelCase__ : int , lowerCamelCase__ : int = 2 , lowerCamelCase__ : int = 2 , **lowerCamelCase__ : Union[str, Any] ) ->Optional[int]: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : str = TFRegNetXLayer if config.layer_type == "x" else TFRegNetYLayer _UpperCAmelCase : List[str] = [ # downsampling is done in the first layer with stride of 2 layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , name="layers.0" ), *[layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , name=F"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def lowerCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : List[str] ) ->List[str]: '''simple docstring''' for layer_module in self.layers: _UpperCAmelCase : Optional[int] = layer_module(lowerCamelCase__ ) return hidden_state class lowerCAmelCase__ ( tf.keras.layers.Layer ): def __init__( self : Dict , lowerCamelCase__ : RegNetConfig , **lowerCamelCase__ : int ) ->Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="stages.0" , ) ) _UpperCAmelCase : Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCamelCase__ , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , depth=lowerCamelCase__ , name=F"""stages.{i+1}""" ) ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = True ) ->TFBaseModelOutputWithNoAttention: '''simple docstring''' _UpperCAmelCase : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase : Optional[Any] = hidden_states + (hidden_state,) _UpperCAmelCase : Dict = stage_module(lowerCamelCase__ ) if output_hidden_states: _UpperCAmelCase : Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase__ , hidden_states=lowerCamelCase__ ) @keras_serializable class lowerCAmelCase__ ( tf.keras.layers.Layer ): lowerCAmelCase : Optional[Any] = RegNetConfig def __init__( self : Union[str, Any] , lowerCamelCase__ : Any , **lowerCamelCase__ : str ) ->int: '''simple docstring''' super().__init__(**lowerCamelCase__ ) _UpperCAmelCase : Union[str, Any] = config _UpperCAmelCase : Union[str, Any] = TFRegNetEmbeddings(lowerCamelCase__ , name="embedder" ) _UpperCAmelCase : Union[str, Any] = TFRegNetEncoder(lowerCamelCase__ , name="encoder" ) _UpperCAmelCase : Union[str, Any] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase__ , name="pooler" ) @unpack_inputs def lowerCAmelCase__ ( self : Any , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : bool = False , ) ->TFBaseModelOutputWithPoolingAndNoAttention: '''simple docstring''' _UpperCAmelCase : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : List[str] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.embedder(lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : str = self.encoder( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : Dict = encoder_outputs[0] _UpperCAmelCase : Dict = self.pooler(lowerCamelCase__ ) # Change to NCHW output format have uniformity in the modules _UpperCAmelCase : Union[str, Any] = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) _UpperCAmelCase : Tuple = tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _UpperCAmelCase : List[str] = tuple([tf.transpose(lowerCamelCase__ , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase__ , pooler_output=lowerCamelCase__ , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): lowerCAmelCase : Tuple = RegNetConfig lowerCAmelCase : Tuple = "regnet" lowerCAmelCase : Union[str, Any] = "pixel_values" @property def lowerCAmelCase__ ( self : Optional[Any] ) ->Optional[int]: '''simple docstring''' return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 2_24, 2_24) , dtype=tf.floataa )} lowerCamelCase__ = r'\n Parameters:\n This model is a Tensorflow\n [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a\n regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and\n behavior.\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights.\n' lowerCamelCase__ = r'\n Args:\n pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConveNextImageProcessor.__call__`] for details.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( "The bare RegNet model outputting raw features without any specific head on top." , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ ): def __init__( self : Any , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : Any , **lowerCamelCase__ : List[str] ) ->Optional[int]: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[Any] = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Optional[bool] = None , lowerCamelCase__ : Any=False , ) ->Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( pixel_values=lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( "\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , UpperCAmelCase__ , ) class lowerCAmelCase__ ( UpperCAmelCase__ , UpperCAmelCase__ ): def __init__( self : str , lowerCamelCase__ : RegNetConfig , *lowerCamelCase__ : List[Any] , **lowerCamelCase__ : Union[str, Any] ) ->Any: '''simple docstring''' super().__init__(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) _UpperCAmelCase : Optional[int] = config.num_labels _UpperCAmelCase : Dict = TFRegNetMainLayer(lowerCamelCase__ , name="regnet" ) # classification head _UpperCAmelCase : str = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="classifier.1" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCAmelCase__ ( self : str , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : tf.Tensor = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : bool = None , lowerCamelCase__ : Dict=False , ) ->Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: '''simple docstring''' _UpperCAmelCase : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase : str = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase : Union[str, Any] = self.regnet( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ , training=lowerCamelCase__ ) _UpperCAmelCase : int = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase : Dict = self.classifier[0](lowerCamelCase__ ) _UpperCAmelCase : str = self.classifier[1](lowerCamelCase__ ) _UpperCAmelCase : Tuple = None if labels is None else self.hf_compute_loss(labels=lowerCamelCase__ , logits=lowerCamelCase__ ) if not return_dict: _UpperCAmelCase : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCamelCase__ , logits=lowerCamelCase__ , hidden_states=outputs.hidden_states )
322
0
'''simple docstring''' from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def __lowerCAmelCase (__lowerCAmelCase ): # A local function to see if a dot lands in the circle. def is_in_circle(__lowerCAmelCase , __lowerCAmelCase ) -> bool: _UpperCAmelCase : Tuple = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle _UpperCAmelCase : str = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(__lowerCAmelCase ) ) # The ratio of the area for circle to square is pi/4. _UpperCAmelCase : List[str] = proportion * 4 print(F"""The estimated value of pi is {pi_estimate}""" ) print(F"""The numpy value of pi is {pi}""" ) print(F"""The total error is {abs(pi - pi_estimate )}""" ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 0.0 , __lowerCAmelCase = 1.0 , ): return mean( function_to_integrate(uniform(__lowerCAmelCase , __lowerCAmelCase ) ) for _ in range(__lowerCAmelCase ) ) * (max_value - min_value) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase = 0.0 , __lowerCAmelCase = 1.0 ): def identity_function(__lowerCAmelCase ) -> float: return x _UpperCAmelCase : Tuple = area_under_curve_estimator( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase : Tuple = (max_value * max_value - min_value * min_value) / 2 print("******************" ) print(F"""Estimating area under y=x where x varies from {min_value} to {max_value}""" ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {expected_value}""" ) print(F"""Total error is {abs(estimated_value - expected_value )}""" ) print("******************" ) def __lowerCAmelCase (__lowerCAmelCase ): def function_to_integrate(__lowerCAmelCase ) -> float: return sqrt(4.0 - x * x ) _UpperCAmelCase : List[str] = area_under_curve_estimator( __lowerCAmelCase , __lowerCAmelCase , 0.0 , 2.0 ) print("******************" ) print("Estimating pi using area_under_curve_estimator" ) print(F"""Estimated value is {estimated_value}""" ) print(F"""Expected value is {pi}""" ) print(F"""Total error is {abs(estimated_value - pi )}""" ) print("******************" ) if __name__ == "__main__": import doctest doctest.testmod()
359
'''simple docstring''' import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def __lowerCAmelCase (__lowerCAmelCase ): if is_torch_version("<" , "2.0.0" ) or not hasattr(__lowerCAmelCase , "_dynamo" ): return False return isinstance(__lowerCAmelCase , torch._dynamo.eval_frame.OptimizedModule ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase = True ): _UpperCAmelCase : Any = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) _UpperCAmelCase : Dict = is_compiled_module(__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : Optional[int] = model _UpperCAmelCase : Any = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = model.module if not keep_fpaa_wrapper: _UpperCAmelCase : List[Any] = getattr(__lowerCAmelCase , "forward" ) _UpperCAmelCase : Dict = model.__dict__.pop("_original_forward" , __lowerCAmelCase ) if original_forward is not None: while hasattr(__lowerCAmelCase , "__wrapped__" ): _UpperCAmelCase : Optional[int] = forward.__wrapped__ if forward == original_forward: break _UpperCAmelCase : Dict = forward if getattr(__lowerCAmelCase , "_converted_to_transformer_engine" , __lowerCAmelCase ): convert_model(__lowerCAmelCase , to_transformer_engine=__lowerCAmelCase ) if is_compiled: _UpperCAmelCase : int = model _UpperCAmelCase : str = compiled_model return model def __lowerCAmelCase (): PartialState().wait_for_everyone() def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): if PartialState().distributed_type == DistributedType.TPU: xm.save(__lowerCAmelCase , __lowerCAmelCase ) elif PartialState().local_process_index == 0: torch.save(__lowerCAmelCase , __lowerCAmelCase ) @contextmanager def __lowerCAmelCase (**__lowerCAmelCase ): for key, value in kwargs.items(): _UpperCAmelCase : str = str(__lowerCAmelCase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def __lowerCAmelCase (__lowerCAmelCase ): if not hasattr(__lowerCAmelCase , "__qualname__" ) and not hasattr(__lowerCAmelCase , "__name__" ): _UpperCAmelCase : List[str] = getattr(__lowerCAmelCase , "__class__" , __lowerCAmelCase ) if hasattr(__lowerCAmelCase , "__qualname__" ): return obj.__qualname__ if hasattr(__lowerCAmelCase , "__name__" ): return obj.__name__ return str(__lowerCAmelCase ) def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): for key, value in source.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Any = destination.setdefault(__lowerCAmelCase , {} ) merge_dicts(__lowerCAmelCase , __lowerCAmelCase ) else: _UpperCAmelCase : Optional[int] = value return destination def __lowerCAmelCase (__lowerCAmelCase = None ): if port is None: _UpperCAmelCase : Tuple = 29_500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
322
0
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase , __lowerCAmelCase ): _UpperCAmelCase : Union[str, Any] = int(__lowerCAmelCase ) # Initialize Result _UpperCAmelCase : Tuple = [] # Traverse through all denomination for denomination in reversed(__lowerCAmelCase ): # Find denominations while int(__lowerCAmelCase ) >= int(__lowerCAmelCase ): total_value -= int(__lowerCAmelCase ) answer.append(__lowerCAmelCase ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": lowerCamelCase__ = [] lowerCamelCase__ = '0' if ( input('Do you want to enter your denominations ? (yY/n): ').strip().lower() == "y" ): lowerCamelCase__ = int(input('Enter the number of denominations you want to add: ').strip()) for i in range(0, n): denominations.append(int(input(F'''Denomination {i}: ''').strip())) lowerCamelCase__ = input('Enter the change you want to make in Indian Currency: ').strip() else: # All denominations of Indian Currency if user does not enter lowerCamelCase__ = [1, 2, 5, 10, 20, 50, 100, 500, 2_000] lowerCamelCase__ = input('Enter the change you want to make: ').strip() if int(value) == 0 or int(value) < 0: print('The total value cannot be zero or negative.') else: print(F'''Following is minimal change for {value}: ''') lowerCamelCase__ = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=' ')
360
'''simple docstring''' def __lowerCAmelCase (__lowerCAmelCase ): if number < 0: raise ValueError("number must not be negative" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
322
0