code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin __A = get_tests_dir("fixtures/test_sentencepiece_bpe.model") class A ( __UpperCAmelCase , unittest.TestCase ): lowerCamelCase : Optional[int] = BartphoTokenizer lowerCamelCase : int = False lowerCamelCase : Tuple = True def A__ ( self ) -> Optional[Any]: '''simple docstring''' super().setUp() lowercase__ = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] lowercase__ = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) lowercase__ = {"""unk_token""": """<unk>"""} lowercase__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(F'''{token} {vocab_tokens[token]}\n''' ) lowercase__ = BartphoTokenizer(lowerCamelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def A__ ( self , **lowerCamelCase__ ) -> Optional[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def A__ ( self , lowerCamelCase__ ) -> List[str]: '''simple docstring''' lowercase__ = """This is a là test""" lowercase__ = """This is a<unk><unk> test""" return input_text, output_text def A__ ( self ) -> Dict: '''simple docstring''' lowercase__ = BartphoTokenizer(lowerCamelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) lowercase__ = """This is a là test""" lowercase__ = """▁This ▁is ▁a ▁l à ▁t est""".split() lowercase__ = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) lowercase__ = tokens + [tokenizer.unk_token] lowercase__ = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ )
164
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer __A = logging.get_logger(__name__) __A = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __A = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } __A = { "yjernite/retribert-base-uncased": 512, } __A = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class A ( __UpperCAmelCase ): lowerCamelCase : int = VOCAB_FILES_NAMES lowerCamelCase : str = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] = PRETRAINED_INIT_CONFIGURATION lowerCamelCase : str = RetriBertTokenizer lowerCamelCase : Optional[int] = ["""input_ids""", """attention_mask"""] def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ) -> Union[str, Any]: '''simple docstring''' super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) lowercase__ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , lowerCamelCase__ ) != do_lower_case or normalizer_state.get("""strip_accents""" , lowerCamelCase__ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , lowerCamelCase__ ) != tokenize_chinese_chars ): lowercase__ = getattr(lowerCamelCase__ , normalizer_state.pop("""type""" ) ) lowercase__ = do_lower_case lowercase__ = strip_accents lowercase__ = tokenize_chinese_chars lowercase__ = normalizer_class(**lowerCamelCase__ ) lowercase__ = do_lower_case def A__ ( self , lowerCamelCase__ , lowerCamelCase__=None ) -> Dict: '''simple docstring''' lowercase__ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> List[int]: '''simple docstring''' lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> Tuple[str]: '''simple docstring''' lowercase__ = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
164
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase: Optional[int] = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase: List[str] = [ """LUKE_PRETRAINED_MODEL_ARCHIVE_LIST""", """LukeForEntityClassification""", """LukeForEntityPairClassification""", """LukeForEntitySpanClassification""", """LukeForMultipleChoice""", """LukeForQuestionAnswering""", """LukeForSequenceClassification""", """LukeForTokenClassification""", """LukeForMaskedLM""", """LukeModel""", """LukePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys UpperCAmelCase: Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
336
"""simple docstring""" import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCAmelCase: Optional[Any] = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCAmelCase: Tuple = [0, 25, 50] UpperCAmelCase: List[Any] = [25, 50, 75] UpperCAmelCase: Optional[int] = fuzz.membership.trimf(X, abca) UpperCAmelCase: Any = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCAmelCase: List[Any] = np.ones(75) UpperCAmelCase: Any = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) UpperCAmelCase: str = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCAmelCase: Optional[int] = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCAmelCase: List[Any] = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCAmelCase: Optional[int] = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCAmelCase: int = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCAmelCase: int = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCAmelCase: List[Any] = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCAmelCase: int = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title("""Young""") plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title("""Middle aged""") plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title("""union""") plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title("""intersection""") plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title("""complement_a""") plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title("""difference a/b""") plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title("""alg_sum""") plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title("""alg_product""") plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title("""bdd_sum""") plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title("""bdd_difference""") plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
336
1
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE : Any = logging.getLogger() def lowercase ( ) ->Optional[int]: """simple docstring""" __snake_case : List[Any] = argparse.ArgumentParser() parser.add_argument('''-f''' ) __snake_case : List[Any] = parser.parse_args() return args.f class _UpperCAmelCase ( A__ ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = logging.StreamHandler(sys.stdout ) logger.addHandler(UpperCamelCase_ ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' __snake_case : Union[str, Any] = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , '''run_glue_deebert.py''' ) with patch.object(UpperCamelCase_ , '''argv''' , UpperCamelCase_ ): __snake_case : Tuple = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(UpperCamelCase_ , 0.666 ) @slow @require_torch_non_multi_gpu def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = ''' --model_type roberta --model_name_or_path roberta-base --task_name MRPC --do_train --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --max_seq_length 128 --per_gpu_eval_batch_size=1 --per_gpu_train_batch_size=8 --learning_rate 2e-4 --num_train_epochs 3 --overwrite_output_dir --seed 42 --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --save_steps 0 --overwrite_cache --eval_after_first_stage '''.split() self.run_and_check(UpperCamelCase_ ) __snake_case : Union[str, Any] = ''' --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --eval_each_highway --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 '''.split() self.run_and_check(UpperCamelCase_ ) __snake_case : str = ''' --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --early_exit_entropy 0.1 --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 '''.split() self.run_and_check(UpperCamelCase_ )
102
'''simple docstring''' from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __snake_case = logging.get_logger(__name__) __snake_case = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __snake_case = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } __snake_case = { '''facebook/blenderbot_small-90M''': 512, } class lowercase ( A__ ): """simple docstring""" _a = VOCAB_FILES_NAMES _a = PRETRAINED_VOCAB_FILES_MAP _a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a = BlenderbotSmallTokenizer def __init__( self , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_="<|endoftext|>" , UpperCamelCase_="<|endoftext|>" , UpperCamelCase_="<|endoftext|>" , UpperCamelCase_=False , UpperCamelCase_=True , **UpperCamelCase_ , ): '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=UpperCamelCase_ , merges=UpperCamelCase_ , add_prefix_space=UpperCamelCase_ , trim_offsets=UpperCamelCase_ , ) , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , **UpperCamelCase_ , ) UpperCamelCase__ :Union[str, Any] = add_prefix_space def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_=None ): '''simple docstring''' UpperCamelCase__ :List[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' UpperCamelCase__ :Optional[int] = [self.sep_token_id] UpperCamelCase__ :Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
97
0
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})''' def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=True ) -> Union[str, Any]: model.train() _lowercase : str = model(lowerCamelCase_ ) _lowercase : List[str] = F.mse_loss(lowerCamelCase_ , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> Optional[int]: set_seed(42 ) _lowercase : Dict = RegressionModel() _lowercase : Tuple = deepcopy(lowerCamelCase_ ) _lowercase : Any = RegressionDataset(length=80 ) _lowercase : Optional[int] = DataLoader(lowerCamelCase_ , batch_size=16 ) model.to(accelerator.device ) if sched: _lowercase : List[str] = AdamW(params=model.parameters() , lr=1e-3 ) _lowercase : Dict = AdamW(params=ddp_model.parameters() , lr=1e-3 ) _lowercase : str = LambdaLR(lowerCamelCase_ , lr_lambda=lambda lowerCamelCase_ : epoch**0.65 ) _lowercase : List[str] = LambdaLR(lowerCamelCase_ , lr_lambda=lambda lowerCamelCase_ : epoch**0.65 ) # Make a copy of `model` if sched: _lowercase : Any = accelerator.prepare(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) else: _lowercase : Union[str, Any] = accelerator.prepare(lowerCamelCase_ , lowerCamelCase_ ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: # Test when on a single CPU or GPU that the context manager does nothing _lowercase : str = get_training_setup(lowerCamelCase_ ) # Use a single batch _lowercase : List[Any] = next(iter(lowerCamelCase_ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model _lowercase : str = accelerator.gather((ddp_input, ddp_target) ) _lowercase : List[str] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase_ ): step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) else: # Sync grads step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) _lowercase : Union[str, Any] = ddp_input[torch.randperm(len(lowerCamelCase_ ) )] def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: # Test on distributed setup that context manager behaves properly _lowercase : Optional[int] = get_training_setup(lowerCamelCase_ ) # Use a single batch _lowercase : Optional[int] = next(iter(lowerCamelCase_ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model _lowercase : Tuple = accelerator.gather((ddp_input, ddp_target) ) _lowercase : str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCamelCase_ ): step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) else: # Sync grads step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) _lowercase : Any = ddp_input[torch.randperm(len(lowerCamelCase_ ) )] def UpperCamelCase_( lowerCamelCase_=False , lowerCamelCase_=False ) -> Dict: _lowercase : Optional[Any] = Accelerator( split_batches=lowerCamelCase_ , dispatch_batches=lowerCamelCase_ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly _lowercase : Optional[int] = get_training_setup(lowerCamelCase_ ) for iteration, batch in enumerate(lowerCamelCase_ ): _lowercase : Any = batch.values() # Gather the distributed inputs and targs for the base model _lowercase : Any = accelerator.gather((ddp_input, ddp_target) ) _lowercase : Dict = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowerCamelCase_ ): step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowerCamelCase_ ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'''Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})''' else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'''Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})''' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) _lowercase : List[str] = ddp_input[torch.randperm(len(lowerCamelCase_ ) )] GradientState._reset_state() def UpperCamelCase_( lowerCamelCase_=False , lowerCamelCase_=False ) -> Union[str, Any]: _lowercase : Any = Accelerator( split_batches=lowerCamelCase_ , dispatch_batches=lowerCamelCase_ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly _lowercase : int = get_training_setup(lowerCamelCase_ , lowerCamelCase_ ) for iteration, batch in enumerate(lowerCamelCase_ ): _lowercase : Any = batch.values() # Gather the distributed inputs and targs for the base model _lowercase : Dict = accelerator.gather((ddp_input, ddp_target) ) _lowercase : int = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowerCamelCase_ )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowerCamelCase_ ): step_model(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'''Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n''' _lowercase : Any = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowerCamelCase_ )) if accelerator.num_processes > 1: check_model_parameters(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def UpperCamelCase_( ) -> List[str]: _lowercase : Optional[Any] = Accelerator() _lowercase : Union[str, Any] = RegressionDataset(length=80 ) _lowercase : str = DataLoader(lowerCamelCase_ , batch_size=16 ) _lowercase : Union[str, Any] = RegressionDataset(length=96 ) _lowercase : Any = DataLoader(lowerCamelCase_ , batch_size=16 ) _lowercase : Optional[int] = accelerator.prepare(lowerCamelCase_ , lowerCamelCase_ ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowerCamelCase_ ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase_ ) if iteration < len(lowerCamelCase_ ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowerCamelCase_ ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCamelCase_ ) if batch_num < len(lowerCamelCase_ ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def UpperCamelCase_( ) -> List[Any]: _lowercase : Dict = Accelerator() _lowercase : Optional[int] = accelerator.state if state.local_process_index == 0: print('**Test `accumulate` gradient accumulation with dataloader break**' ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print('**Test NOOP `no_sync` context manager**' ) test_noop_sync(lowerCamelCase_ ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print('**Test Distributed `no_sync` context manager**' ) test_distributed_sync(lowerCamelCase_ ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation, ' , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation(lowerCamelCase_ , lowerCamelCase_ ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version('<' , '2.0' ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation with optimizer and scheduler, ' , '`split_batches=False`, `dispatch_batches=False`**' , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation with optimizer and scheduler, ' , F'''`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**''' , ) test_gradient_accumulation_with_opt_and_scheduler(lowerCamelCase_ , lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
356
from collections import defaultdict def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Optional[Any] = 1 _lowercase : Union[str, Any] = True for v in tree[start]: if v not in visited: ret += dfs(lowerCamelCase_ ) if ret % 2 == 0: cuts.append(lowerCamelCase_ ) return ret def UpperCamelCase_( ) -> Optional[Any]: dfs(1 ) if __name__ == "__main__": SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[str] = 10, 9 SCREAMING_SNAKE_CASE : List[Any] = defaultdict(list) SCREAMING_SNAKE_CASE : dict[int, bool] = {} SCREAMING_SNAKE_CASE : list[int] = [] SCREAMING_SNAKE_CASE : Optional[int] = 0 SCREAMING_SNAKE_CASE : Any = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
84
0
'''simple docstring''' from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def _lowercase ( __A = "isbn/0140328726" ): '''simple docstring''' __UpperCamelCase = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: __UpperCamelCase = f"{olid} is not a valid Open Library olid" raise ValueError(__A ) return requests.get(f"https://openlibrary.org/{new_olid}.json" ).json() def _lowercase ( __A ): '''simple docstring''' __UpperCamelCase = { """title""": """Title""", """publish_date""": """Publish date""", """authors""": """Authors""", """number_of_pages""": """Number of pages:""", """first_sentence""": """First sentence""", """isbn_10""": """ISBN (10)""", """isbn_13""": """ISBN (13)""", } __UpperCamelCase = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} __UpperCamelCase = [ get_openlibrary_data(author["""key"""] )["""name"""] for author in data["""Authors"""] ] __UpperCamelCase = data["""First sentence"""]["""value"""] for key, value in data.items(): if isinstance(__A ,__A ): __UpperCamelCase = """, """.join(__A ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: a__ : Tuple = input('\nEnter the ISBN code to search (or \'quit\' to stop): ').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (1_0, 1_3) or not isbn.isdigit(): print(f'''Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.''') continue print(f'''\nSearching Open Library for ISBN: {isbn}...\n''') try: a__ : str = summarize_book(get_openlibrary_data(f'''isbn/{isbn}''')) print('\n'.join(f'''{key}: {value}''' for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(f'''Sorry, there are no results for ISBN: {isbn}.''')
349
'''simple docstring''' import pytest a__ : List[str] = '__dummy_dataset1__' a__ : Optional[int] = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def _lowercase ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def _lowercase ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def _lowercase ( __A ,__A ,__A ): '''simple docstring''' __UpperCamelCase = dataset_loading_script_name __UpperCamelCase = tmp_path / """datasets""" / script_name script_dir.mkdir(parents=__A ) __UpperCamelCase = script_dir / f"{script_name}.py" with open(__A ,"""w""" ) as f: f.write(__A ) return str(__A )
349
1
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class UpperCamelCase__ ( lowercase_, lowercase_, unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = IFInpaintingPipeline _SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} _SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _SCREAMING_SNAKE_CASE = PipelineTesterMixin.required_optional_params - {"""latents"""} def SCREAMING_SNAKE_CASE__ ( self : Dict ): return self._get_dummy_components() def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : Dict=0 ): if str(SCREAMING_SNAKE_CASE_ ).startswith('mps' ): lowerCAmelCase_ : Dict = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) else: lowerCAmelCase_ : Dict = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : Any = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(SCREAMING_SNAKE_CASE_ ) ).to(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : Optional[int] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(SCREAMING_SNAKE_CASE_ ) ).to(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : int = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def SCREAMING_SNAKE_CASE__ ( self : Any ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): self._test_save_load_local() def SCREAMING_SNAKE_CASE__ ( self : Dict ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
361
"""simple docstring""" from __future__ import annotations def UpperCamelCase_ ( lowerCAmelCase__ : int ) -> list[int]: """simple docstring""" lowerCAmelCase_ : Any = 2 lowerCAmelCase_ : List[Any] = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(lowerCAmelCase__ ) if n > 1: factors.append(lowerCAmelCase__ ) return factors if __name__ == "__main__": import doctest doctest.testmod()
289
0
from typing import List from .keymap import KEYMAP, get_character def lowerCamelCase__ ( _A ): '''simple docstring''' def decorator(_A ): snake_case_ = getattr(_A , "handle_key" , [] ) handle += [key] setattr(_A , "handle_key" , _A ) return func return decorator def lowerCamelCase__ ( *_A ): '''simple docstring''' def decorator(_A ): snake_case_ = getattr(_A , "handle_key" , [] ) handle += keys setattr(_A , "handle_key" , _A ) return func return decorator class UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __new__( cls : List[Any] , __lowercase : Tuple , __lowercase : Any , __lowercase : Tuple ): """simple docstring""" snake_case_ = super().__new__(cls , __lowercase , __lowercase , __lowercase ) if not hasattr(__lowercase , "key_handler" ): setattr(__lowercase , "key_handler" , {} ) setattr(__lowercase , "handle_input" , KeyHandler.handle_input ) for value in attrs.values(): snake_case_ = getattr(__lowercase , "handle_key" , [] ) for key in handled_keys: snake_case_ = value return new_cls @staticmethod def snake_case__ ( cls : str ): """simple docstring""" snake_case_ = get_character() if char != KEYMAP["undefined"]: snake_case_ = ord(__lowercase ) snake_case_ = cls.key_handler.get(__lowercase ) if handler: snake_case_ = char return handler(cls ) else: return None def lowerCamelCase__ ( cls ): '''simple docstring''' return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
187
def lowerCamelCase__ ( _A ): '''simple docstring''' if num <= 0: raise ValueError("Input must be a positive integer" ) snake_case_ = [True] * (num + 1) snake_case_ = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , _A ): snake_case_ = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() lowercase__ : Tuple = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
187
1
'''simple docstring''' def __lowerCamelCase ( lowerCAmelCase_ ) -> str: _a : Optional[int] = 0 _a : str = len(lowerCAmelCase_ ) for i in range(n - 1 ): for j in range(i + 1 , lowerCAmelCase_ ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def __lowerCamelCase ( lowerCAmelCase_ ) -> Dict: if len(lowerCAmelCase_ ) <= 1: return arr, 0 _a : Tuple = len(lowerCAmelCase_ ) // 2 _a : Any = arr[0:mid] _a : List[Any] = arr[mid:] _a , _a : Optional[int] = count_inversions_recursive(lowerCAmelCase_ ) _a , _a : int = count_inversions_recursive(lowerCAmelCase_ ) _a , _a : Optional[Any] = _count_cross_inversions(lowerCAmelCase_ , lowerCAmelCase_ ) _a : Optional[int] = inversion_p + inversions_q + cross_inversions return c, num_inversions def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Union[str, Any]: _a : List[str] = [] _a : Dict = 0 while i < len(lowerCAmelCase_ ) and j < len(lowerCAmelCase_ ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(lowerCAmelCase_ ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(lowerCAmelCase_ ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def __lowerCamelCase ( ) -> Tuple: _a : Union[str, Any] = [10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) _a : str = count_inversions_bf(lowerCAmelCase_ ) _a , _a : Union[str, Any] = count_inversions_recursive(lowerCAmelCase_ ) assert num_inversions_bf == num_inversions_recursive == 8 print('number of inversions = ' , lowerCAmelCase_ ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() _a : int = count_inversions_bf(lowerCAmelCase_ ) _a , _a : Tuple = count_inversions_recursive(lowerCAmelCase_ ) assert num_inversions_bf == num_inversions_recursive == 0 print('number of inversions = ' , lowerCAmelCase_ ) # an empty list should also have zero inversions _a : int = [] _a : int = count_inversions_bf(lowerCAmelCase_ ) _a , _a : Union[str, Any] = count_inversions_recursive(lowerCAmelCase_ ) assert num_inversions_bf == num_inversions_recursive == 0 print('number of inversions = ' , lowerCAmelCase_ ) if __name__ == "__main__": main()
107
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class __magic_name__ : lowerCAmelCase : str = field( metadata={'help': 'The output directory where the model will be written.'} , ) lowerCAmelCase : str = field( metadata={ 'help': ( 'The encoder model checkpoint for weights initialization.' 'Don\'t set if you want to train an encoder model from scratch.' ) } , ) lowerCAmelCase : str = field( metadata={ 'help': ( 'The decoder model checkpoint for weights initialization.' 'Don\'t set if you want to train a decoder model from scratch.' ) } , ) lowerCAmelCase : Optional[str] = field( default=_UpperCamelCase , metadata={'help': 'Pretrained encoder config name or path if not the same as encoder_model_name'} ) lowerCAmelCase : Optional[str] = field( default=_UpperCamelCase , metadata={'help': 'Pretrained decoder config name or path if not the same as decoder_model_name'} ) def __lowerCamelCase ( ) -> Union[str, Any]: _a : Any = HfArgumentParser((ModelArguments,) ) ((_a) , ) : Dict = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _a : Optional[Any] = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _a : Optional[Any] = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _a : List[str] = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _a : Optional[int] = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _a : List[Any] = True _a : int = True _a : Any = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=lowerCAmelCase_ , decoder_config=lowerCAmelCase_ , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _a : List[str] = decoder_config.decoder_start_token_id _a : Optional[int] = decoder_config.pad_token_id if decoder_start_token_id is None: _a : Tuple = decoder_config.bos_token_id if pad_token_id is None: _a : List[Any] = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _a : Any = decoder_config.eos_token_id _a : Tuple = decoder_start_token_id _a : Any = pad_token_id _a : Dict = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _a : Dict = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _a : int = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
107
1
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union __SCREAMING_SNAKE_CASE : str = TypeVar("""T""") __SCREAMING_SNAKE_CASE : int = Union[List[T], Tuple[T, ...]] __SCREAMING_SNAKE_CASE : Any = Union[T, List[T], Dict[str, T]] __SCREAMING_SNAKE_CASE : str = Union[str, bytes, os.PathLike]
31
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowercase__ ( _snake_case ): '''simple docstring''' A_ : UNetaDModel A_ : ScoreSdeVeScheduler def __init__( self , __snake_case , __snake_case ): super().__init__() self.register_modules(unet=__snake_case , scheduler=__snake_case ) @torch.no_grad() def __call__( self , __snake_case = 1 , __snake_case = 2000 , __snake_case = None , __snake_case = "pil" , __snake_case = True , **__snake_case , ): _SCREAMING_SNAKE_CASE : Optional[int] = self.unet.config.sample_size _SCREAMING_SNAKE_CASE : Optional[Any] = (batch_size, 3, img_size, img_size) _SCREAMING_SNAKE_CASE : Union[str, Any] = self.unet _SCREAMING_SNAKE_CASE : Optional[int] = randn_tensor(__snake_case , generator=__snake_case ) * self.scheduler.init_noise_sigma _SCREAMING_SNAKE_CASE : Any = sample.to(self.device ) self.scheduler.set_timesteps(__snake_case ) self.scheduler.set_sigmas(__snake_case ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): _SCREAMING_SNAKE_CASE : List[Any] = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): _SCREAMING_SNAKE_CASE : Any = self.unet(__snake_case , __snake_case ).sample _SCREAMING_SNAKE_CASE : Any = self.scheduler.step_correct(__snake_case , __snake_case , generator=__snake_case ).prev_sample # prediction step _SCREAMING_SNAKE_CASE : Union[str, Any] = model(__snake_case , __snake_case ).sample _SCREAMING_SNAKE_CASE : Tuple = self.scheduler.step_pred(__snake_case , __snake_case , __snake_case , generator=__snake_case ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : int = output.prev_sample, output.prev_sample_mean _SCREAMING_SNAKE_CASE : Tuple = sample_mean.clamp(0 , 1 ) _SCREAMING_SNAKE_CASE : str = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _SCREAMING_SNAKE_CASE : Optional[int] = self.numpy_to_pil(__snake_case ) if not return_dict: return (sample,) return ImagePipelineOutput(images=__snake_case )
200
0
import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCamelCase_( snake_case__: Optional[int] ) -> Union[str, Any]: # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0X4E00 and cp <= 0X9FFF) or (cp >= 0X3400 and cp <= 0X4DBF) # or (cp >= 0X20000 and cp <= 0X2A6DF) # or (cp >= 0X2A700 and cp <= 0X2B73F) # or (cp >= 0X2B740 and cp <= 0X2B81F) # or (cp >= 0X2B820 and cp <= 0X2CEAF) # or (cp >= 0XF900 and cp <= 0XFAFF) or (cp >= 0X2F800 and cp <= 0X2FA1F) # ): # return True return False def UpperCamelCase_( snake_case__: str ) -> str: # word like '180' or '身高' or '神' for char in word: UpperCAmelCase__ = ord(_a ) if not _is_chinese_char(_a ): return 0 return 1 def UpperCamelCase_( snake_case__: List[str] ) -> int: UpperCAmelCase__ = set() for token in tokens: UpperCAmelCase__ = len(_a ) > 1 and is_chinese(_a ) if chinese_word: word_set.add(_a ) UpperCAmelCase__ = list(_a ) return word_list def UpperCamelCase_( snake_case__: List[str] , snake_case__: set() ) -> List[str]: if not chinese_word_set: return bert_tokens UpperCAmelCase__ = max([len(_a ) for w in chinese_word_set] ) UpperCAmelCase__ = bert_tokens UpperCAmelCase__ = 0, len(_a ) while start < end: UpperCAmelCase__ = True if is_chinese(bert_word[start] ): UpperCAmelCase__ = min(end - start , _a ) for i in range(_a , 1 , -1 ): UpperCAmelCase__ = "".join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): UpperCAmelCase__ = "##" + bert_word[j] UpperCAmelCase__ = start + i UpperCAmelCase__ = False break if single_word: start += 1 return bert_word def UpperCamelCase_( snake_case__: List[str] , snake_case__: LTP , snake_case__: BertTokenizer ) -> Optional[Any]: UpperCAmelCase__ = [] for i in range(0 , len(_a ) , 1_00 ): UpperCAmelCase__ = ltp_tokenizer.seg(lines[i : i + 1_00] )[0] UpperCAmelCase__ = [get_chinese_word(_a ) for r in res] ltp_res.extend(_a ) assert len(_a ) == len(_a ) UpperCAmelCase__ = [] for i in range(0 , len(_a ) , 1_00 ): UpperCAmelCase__ = bert_tokenizer(lines[i : i + 1_00] , add_special_tokens=_a , truncation=_a , max_length=5_12 ) bert_res.extend(res['input_ids'] ) assert len(_a ) == len(_a ) UpperCAmelCase__ = [] for input_ids, chinese_word in zip(_a , _a ): UpperCAmelCase__ = [] for id in input_ids: UpperCAmelCase__ = bert_tokenizer._convert_id_to_token(_a ) input_tokens.append(_a ) UpperCAmelCase__ = add_sub_symbol(_a , _a ) UpperCAmelCase__ = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(_a ): if token[:2] == "##": UpperCAmelCase__ = token[2:] # save chinese tokens' pos if len(_a ) == 1 and _is_chinese_char(ord(_a ) ): ref_id.append(_a ) ref_ids.append(_a ) assert len(_a ) == len(_a ) return ref_ids def UpperCamelCase_( snake_case__: int ) -> str: # For Chinese (Ro)Bert, the best result is from : RoBERTa-wwm-ext (https://github.com/ymcui/Chinese-BERT-wwm) # If we want to fine-tune these model, we have to use same tokenizer : LTP (https://github.com/HIT-SCIR/ltp) with open(args.file_name , 'r' , encoding='utf-8' ) as f: UpperCAmelCase__ = f.readlines() UpperCAmelCase__ = [line.strip() for line in data if len(_a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' UpperCAmelCase__ = LTP(args.ltp ) # faster in GPU device UpperCAmelCase__ = BertTokenizer.from_pretrained(args.bert ) UpperCAmelCase__ = prepare_ref(_a , _a , _a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: UpperCAmelCase__ = [json.dumps(_a ) + "\n" for ref in ref_ids] f.writelines(_a ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser(description='''prepare_chinese_ref''') parser.add_argument( '''--file_name''', type=str, default='''./resources/chinese-demo.txt''', help='''file need process, same as training data in lm''', ) parser.add_argument( '''--ltp''', type=str, default='''./resources/ltp''', help='''resources for LTP tokenizer, usually a path''' ) parser.add_argument('''--bert''', type=str, default='''./resources/robert''', help='''resources for Bert tokenizer''') parser.add_argument('''--save_path''', type=str, default='''./resources/ref.txt''', help='''path to save res''') _UpperCamelCase = parser.parse_args() main(args)
364
from dataclasses import dataclass, field from typing import Optional @dataclass class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Model name or path of model to be trained."""} ) __SCREAMING_SNAKE_CASE = field( default="""./""" , metadata={"""help""": """Save dir where model repo is cloned and models updates are saved to."""} ) __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot-clean-train""" , metadata={"""help""": """Name or path of training dataset."""} ) __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot-clean-valid""" , metadata={"""help""": """Name or path of validation dataset."""} ) __SCREAMING_SNAKE_CASE = field(default=2 , metadata={"""help""": """Batch size for training."""} ) __SCREAMING_SNAKE_CASE = field(default=2 , metadata={"""help""": """Batch size for evaluation."""} ) __SCREAMING_SNAKE_CASE = field(default=0.1 , metadata={"""help""": """Value of weight decay."""} ) __SCREAMING_SNAKE_CASE = field( default=10000 , metadata={"""help""": """Size of buffer used to shuffle streaming dataset."""} ) __SCREAMING_SNAKE_CASE = field(default=2E-4 , metadata={"""help""": """Learning rate fo training."""} ) __SCREAMING_SNAKE_CASE = field(default="""cosine""" , metadata={"""help""": """Learning rate."""} ) __SCREAMING_SNAKE_CASE = field( default=750 , metadata={"""help""": """Number of warmup steps in the learning rate schedule."""} ) __SCREAMING_SNAKE_CASE = field( default=16 , metadata={"""help""": """Number of gradient accumulation steps."""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Use gradient checkpointing to reduce memory footprint."""} ) __SCREAMING_SNAKE_CASE = field(default=50000 , metadata={"""help""": """Maximum number of training steps."""} ) __SCREAMING_SNAKE_CASE = field( default=-1 , metadata={"""help""": """Maximum number of evaluation steps. If -1 the full dataset is evaluated."""} ) __SCREAMING_SNAKE_CASE = field(default=1024 , metadata={"""help""": """Sequence lengths used for training."""} ) __SCREAMING_SNAKE_CASE = field(default=1 , metadata={"""help""": """Training seed."""} ) __SCREAMING_SNAKE_CASE = field( default=1024 , metadata={"""help""": """Interval to save checkpoints. Measured as number of forward passes not training steps."""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """States path if the training should continue from a checkpoint folder."""} ) __SCREAMING_SNAKE_CASE = field(default=_UpperCamelCase , metadata={"""help""": """If True the data is pretokenized."""} ) @dataclass class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Model name or path of model to be evaluated."""} ) __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot-clean-valid""" , metadata={"""help""": """Name or path of validation dataset."""} ) __SCREAMING_SNAKE_CASE = field(default=2 , metadata={"""help""": """Batch size used for evaluation."""} ) __SCREAMING_SNAKE_CASE = field( default=-1 , metadata={"""help""": """Maximum number of evaluation steps. If -1 the full dataset is evaluated."""} ) __SCREAMING_SNAKE_CASE = field(default=1024 , metadata={"""help""": """Length of sequences to be evaluated."""} ) __SCREAMING_SNAKE_CASE = field(default=1 , metadata={"""help""": """Random seed used for evaluation."""} ) @dataclass class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Model name or path of model to be evaluated."""} ) __SCREAMING_SNAKE_CASE = field(default=_UpperCamelCase , metadata={"""help""": """Number of workers used for code evaluation."""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """The number of human-eval tasks to run. If not included all tasks are evaluated."""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Sample from the language model's output distribution."""} ) __SCREAMING_SNAKE_CASE = field(default=0.2 , metadata={"""help""": """Sampling temperature used for generation."""} ) __SCREAMING_SNAKE_CASE = field(default=256 , metadata={"""help""": """Maximum number of newly generated tokens."""} ) __SCREAMING_SNAKE_CASE = field(default=0 , metadata={"""help""": """Top-k parameter used for generation."""} ) __SCREAMING_SNAKE_CASE = field(default=0.95 , metadata={"""help""": """Top-p parameter used for nucleus sampling."""} ) __SCREAMING_SNAKE_CASE = field(default=10 , metadata={"""help""": """Number of generations to run in parallel."""} ) __SCREAMING_SNAKE_CASE = field( default=200 , metadata={"""help""": """Number of completions to generate for each sample."""} ) __SCREAMING_SNAKE_CASE = field(default=1 , metadata={"""help""": """Random seed used for evaluation."""} ) __SCREAMING_SNAKE_CASE = field( default="""eval_results.json""" , metadata={"""help""": """Random seed used for evaluation."""} ) __SCREAMING_SNAKE_CASE = field( default="""0""" , metadata={"""help""": """Allow `code_eval` to execute Python code on machine"""} ) __SCREAMING_SNAKE_CASE = field( default=-1 , metadata={ """help""": ( """Determine which device to run the `text-generation` Pipeline on. -1 is CPU and any zero or positive""" """ number corresponds to which GPU device id to run on.""" ) } , ) @dataclass class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": """The number of CPU cores to use for parallel preprocessing. Default uses the maximum available.""" } , ) __SCREAMING_SNAKE_CASE = field( default="""transformersbook/codeparrot""" , metadata={"""help""": """Folder or name of dataset to process."""} ) __SCREAMING_SNAKE_CASE = field( default="""codeparrot-clean""" , metadata={"""help""": """Folder to save processed processed dataset."""} ) __SCREAMING_SNAKE_CASE = field( default=100000 , metadata={"""help""": """Number of files to save per JSON output file."""} ) __SCREAMING_SNAKE_CASE = field(default="""content""" , metadata={"""help""": """Column containing text data to process."""} ) __SCREAMING_SNAKE_CASE = field( default=1000 , metadata={"""help""": """Maximum line length in file, otherwise file is filtered."""} ) __SCREAMING_SNAKE_CASE = field( default=100 , metadata={"""help""": """Maximum mean line length in file, otherwise file is filtered."""} ) __SCREAMING_SNAKE_CASE = field( default=0.25 , metadata={"""help""": """Maximum fraction of non-alphanumeric characters, otherwise file is filtered."""} ) __SCREAMING_SNAKE_CASE = field( default=1.5 , metadata={"""help""": """Minimum character token ratio for the file, otherwise file is filtered."""} ) __SCREAMING_SNAKE_CASE = field( default=0.7 , metadata={"""help""": """Probability for filtering config, test and uncommon files."""} ) __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Name or path to the tokenizer."""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """If True, near-duplicate samples are removed."""} ) __SCREAMING_SNAKE_CASE = field( default=0.85 , metadata={"""help""": """Jaccard threshold for near-duplicate samples."""} ) @dataclass class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = field( default="""gpt2""" , metadata={"""help""": """Base tokenizer to build new tokenizer from."""} ) __SCREAMING_SNAKE_CASE = field( default="""transformersbook/codeparrot-train""" , metadata={"""help""": """Dataset to train tokenizer on."""} ) __SCREAMING_SNAKE_CASE = field(default="""content""" , metadata={"""help""": """Column containing text data to process."""} ) __SCREAMING_SNAKE_CASE = field(default=200000 , metadata={"""help""": """Number of examples to train tokenizer on."""} ) __SCREAMING_SNAKE_CASE = field( default=32768 , metadata={"""help""": """Number of examples to train the tokenizer on."""} ) __SCREAMING_SNAKE_CASE = field(default="""codeparrot""" , metadata={"""help""": """Name of new tokenizer."""} ) __SCREAMING_SNAKE_CASE = field(default=_UpperCamelCase , metadata={"""help""": """Push saved tokenizer to the hub."""} ) @dataclass class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Name or path to the tokenizer."""} ) __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot-clean-train""" , metadata={"""help""": """Name or path to the dataset to pretokenize."""} ) __SCREAMING_SNAKE_CASE = field( default="""tokenized-codeparrot-train""" , metadata={"""help""": """Repo name of the pretokenized data."""} ) __SCREAMING_SNAKE_CASE = field(default=_UpperCamelCase , metadata={"""help""": """Number of workers used for code evaluation."""} ) @dataclass class lowercase : '''simple docstring''' __SCREAMING_SNAKE_CASE = field( default="""gpt2-large""" , metadata={"""help""": """Configuration to use for model initialization."""} ) __SCREAMING_SNAKE_CASE = field( default="""codeparrot/codeparrot""" , metadata={"""help""": """Tokenizer attached to model."""} ) __SCREAMING_SNAKE_CASE = field(default="""codeparrot""" , metadata={"""help""": """Name of the created model."""} ) __SCREAMING_SNAKE_CASE = field(default=_UpperCamelCase , metadata={"""help""": """Push saved tokenizer to the hub."""} )
335
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def __UpperCAmelCase ( A : Optional[int] ) -> Dict: UpperCAmelCase_ : Optional[int] = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class snake_case__ ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase): a_ = StableDiffusionLatentUpscalePipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "height", "width", "cross_attention_kwargs", "negative_prompt_embeds", "prompt_embeds", } a_ = PipelineTesterMixin.required_optional_params - {"num_images_per_prompt"} a_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a_ = frozenset( []) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a_ = frozenset([]) a_ = True @property def A ( self : Union[str, Any] ) -> Dict: UpperCAmelCase_ : Union[str, Any] = 1 UpperCAmelCase_ : Optional[int] = 4 UpperCAmelCase_ : List[Any] = (16, 16) UpperCAmelCase_ : Any = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_A ) return image def A ( self : Optional[Any] ) -> Optional[int]: torch.manual_seed(0 ) UpperCAmelCase_ : Optional[int] = UNetaDConditionModel( act_fn='''gelu''' , attention_head_dim=8 , norm_num_groups=_A , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=1_60 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( '''KDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', ) , in_channels=8 , mid_block_type=_A , only_cross_attention=_A , out_channels=5 , resnet_time_scale_shift='''scale_shift''' , time_embedding_type='''fourier''' , timestep_post_act='''gelu''' , up_block_types=('''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KUpBlock2D''') , ) UpperCAmelCase_ : Optional[int] = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', ] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) UpperCAmelCase_ : Optional[Any] = EulerDiscreteScheduler(prediction_type='''sample''' ) UpperCAmelCase_ : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''quick_gelu''' , projection_dim=5_12 , ) UpperCAmelCase_ : List[str] = CLIPTextModel(_A ) UpperCAmelCase_ : List[str] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) UpperCAmelCase_ : Union[str, Any] = { '''unet''': model.eval(), '''vae''': vae.eval(), '''scheduler''': scheduler, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def A ( self : Any , _A : Tuple , _A : Any=0 ) -> Optional[Any]: if str(_A ).startswith('''mps''' ): UpperCAmelCase_ : Union[str, Any] = torch.manual_seed(_A ) else: UpperCAmelCase_ : Any = torch.Generator(device=_A ).manual_seed(_A ) UpperCAmelCase_ : List[str] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': self.dummy_image.cpu(), '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def A ( self : List[Any] ) -> Dict: UpperCAmelCase_ : Optional[Any] = '''cpu''' UpperCAmelCase_ : Optional[Any] = self.get_dummy_components() UpperCAmelCase_ : Optional[int] = self.pipeline_class(**_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) UpperCAmelCase_ : List[Any] = self.get_dummy_inputs(_A ) UpperCAmelCase_ : Dict = pipe(**_A ).images UpperCAmelCase_ : int = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 2_56, 2_56, 3) ) UpperCAmelCase_ : List[Any] = np.array( [0.47_222_412, 0.41_921_633, 0.44_717_434, 0.46_874_192, 0.42_588_258, 0.46_150_726, 0.4_677_534, 0.45_583_832, 0.48_579_055] ) UpperCAmelCase_ : List[str] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_A , 1e-3 ) def A ( self : str ) -> Optional[int]: super().test_attention_slicing_forward_pass(expected_max_diff=7e-3 ) def A ( self : str ) -> Any: super().test_cpu_offload_forward_pass(expected_max_diff=3e-3 ) def A ( self : Union[str, Any] ) -> Union[str, Any]: super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def A ( self : Any ) -> Optional[Any]: super().test_inference_batch_single_identical(expected_max_diff=7e-3 ) def A ( self : List[Any] ) -> str: super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3 ) def A ( self : List[str] ) -> Optional[Any]: super().test_save_load_local(expected_max_difference=3e-3 ) def A ( self : Any ) -> int: super().test_save_load_optional_components(expected_max_difference=3e-3 ) def A ( self : Any ) -> Any: UpperCAmelCase_ : List[Any] = [ '''DDIMScheduler''', '''DDPMScheduler''', '''PNDMScheduler''', '''HeunDiscreteScheduler''', '''EulerAncestralDiscreteScheduler''', '''KDPM2DiscreteScheduler''', '''KDPM2AncestralDiscreteScheduler''', '''DPMSolverSDEScheduler''', ] UpperCAmelCase_ : Optional[Any] = self.get_dummy_components() UpperCAmelCase_ : Dict = self.pipeline_class(**_A ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) UpperCAmelCase_ : List[str] = self.get_dummy_inputs(_A ) UpperCAmelCase_ : int = 2 UpperCAmelCase_ : Optional[Any] = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue UpperCAmelCase_ : List[Any] = getattr(_A , scheduler_enum.name ) UpperCAmelCase_ : List[Any] = scheduler_cls.from_config(pipe.scheduler.config ) UpperCAmelCase_ : Any = pipe(**_A )[0] outputs.append(_A ) assert check_same_shape(_A ) @require_torch_gpu @slow class snake_case__ ( unittest.TestCase): def A ( self : List[Any] ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Optional[int] ) -> Dict: UpperCAmelCase_ : Any = torch.manual_seed(33 ) UpperCAmelCase_ : Union[str, Any] = StableDiffusionPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''' , torch_dtype=torch.floataa ) pipe.to('''cuda''' ) UpperCAmelCase_ : str = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) UpperCAmelCase_ : Union[str, Any] = '''a photo of an astronaut high resolution, unreal engine, ultra realistic''' UpperCAmelCase_ : Tuple = pipe(_A , generator=_A , output_type='''latent''' ).images UpperCAmelCase_ : Optional[int] = upscaler( prompt=_A , image=_A , num_inference_steps=20 , guidance_scale=0 , generator=_A , output_type='''np''' , ).images[0] UpperCAmelCase_ : Optional[int] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy''' ) assert np.abs((expected_image - image).mean() ) < 5e-2 def A ( self : int ) -> str: UpperCAmelCase_ : Optional[int] = torch.manual_seed(33 ) UpperCAmelCase_ : Union[str, Any] = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) UpperCAmelCase_ : List[str] = '''the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas''' UpperCAmelCase_ : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png''' ) UpperCAmelCase_ : Union[str, Any] = upscaler( prompt=_A , image=_A , num_inference_steps=20 , guidance_scale=0 , generator=_A , output_type='''np''' , ).images[0] UpperCAmelCase_ : Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy''' ) assert np.abs((expected_image - image).max() ) < 5e-2
304
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class snake_case__ ( UpperCamelCase , UpperCamelCase , unittest.TestCase): a_ = StableDiffusionDiffEditPipeline a_ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"height", "width", "image"} | {"image_latents"} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"image"} | {"image_latents"} a_ = frozenset( []) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a_ = frozenset([]) def A ( self : Tuple ) -> Optional[Any]: torch.manual_seed(0 ) UpperCAmelCase_ : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=_A , ) UpperCAmelCase_ : Optional[Any] = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=_A , set_alpha_to_one=_A , ) UpperCAmelCase_ : Optional[int] = DDIMInverseScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=_A , set_alpha_to_zero=_A , ) torch.manual_seed(0 ) UpperCAmelCase_ : List[str] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) UpperCAmelCase_ : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''gelu''' , projection_dim=5_12 , ) UpperCAmelCase_ : Union[str, Any] = CLIPTextModel(_A ) UpperCAmelCase_ : List[Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) UpperCAmelCase_ : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''inverse_scheduler''': inverse_scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def A ( self : str , _A : List[str] , _A : Any=0 ) -> str: UpperCAmelCase_ : Optional[Any] = floats_tensor((1, 16, 16) , rng=random.Random(_A ) ).to(_A ) UpperCAmelCase_ : Dict = floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(_A ) ).to(_A ) if str(_A ).startswith('''mps''' ): UpperCAmelCase_ : Any = torch.manual_seed(_A ) else: UpperCAmelCase_ : Tuple = torch.Generator(device=_A ).manual_seed(_A ) UpperCAmelCase_ : str = { '''prompt''': '''a dog and a newt''', '''mask_image''': mask, '''image_latents''': latents, '''generator''': generator, '''num_inference_steps''': 2, '''inpaint_strength''': 1.0, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : Tuple , _A : Optional[Any] , _A : Optional[Any]=0 ) -> List[str]: UpperCAmelCase_ : Union[str, Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A ) ).to(_A ) UpperCAmelCase_ : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCAmelCase_ : int = Image.fromarray(np.uinta(_A ) ).convert('''RGB''' ) if str(_A ).startswith('''mps''' ): UpperCAmelCase_ : Dict = torch.manual_seed(_A ) else: UpperCAmelCase_ : Any = torch.Generator(device=_A ).manual_seed(_A ) UpperCAmelCase_ : Optional[Any] = { '''image''': image, '''source_prompt''': '''a cat and a frog''', '''target_prompt''': '''a dog and a newt''', '''generator''': generator, '''num_inference_steps''': 2, '''num_maps_per_mask''': 2, '''mask_encode_strength''': 1.0, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : int , _A : Tuple , _A : List[str]=0 ) -> Any: UpperCAmelCase_ : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A ) ).to(_A ) UpperCAmelCase_ : List[str] = image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCAmelCase_ : Optional[int] = Image.fromarray(np.uinta(_A ) ).convert('''RGB''' ) if str(_A ).startswith('''mps''' ): UpperCAmelCase_ : Optional[int] = torch.manual_seed(_A ) else: UpperCAmelCase_ : Tuple = torch.Generator(device=_A ).manual_seed(_A ) UpperCAmelCase_ : Optional[int] = { '''image''': image, '''prompt''': '''a cat and a frog''', '''generator''': generator, '''num_inference_steps''': 2, '''inpaint_strength''': 1.0, '''guidance_scale''': 6.0, '''decode_latents''': True, '''output_type''': '''numpy''', } return inputs def A ( self : List[str] ) -> Optional[Any]: if not hasattr(self.pipeline_class , '''_optional_components''' ): return UpperCAmelCase_ : str = self.get_dummy_components() UpperCAmelCase_ : Any = self.pipeline_class(**_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(_A , _A , _A ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) UpperCAmelCase_ : List[str] = self.get_dummy_inputs(_A ) UpperCAmelCase_ : str = pipe(**_A )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_A ) UpperCAmelCase_ : Any = self.pipeline_class.from_pretrained(_A ) pipe_loaded.to(_A ) pipe_loaded.set_progress_bar_config(disable=_A ) for optional_component in pipe._optional_components: self.assertTrue( getattr(_A , _A ) is None , F"`{optional_component}` did not stay set to None after loading." , ) UpperCAmelCase_ : Tuple = self.get_dummy_inputs(_A ) UpperCAmelCase_ : List[Any] = pipe_loaded(**_A )[0] UpperCAmelCase_ : Any = np.abs(output - output_loaded ).max() self.assertLess(_A , 1e-4 ) def A ( self : Tuple ) -> int: UpperCAmelCase_ : Optional[Any] = '''cpu''' UpperCAmelCase_ : Any = self.get_dummy_components() UpperCAmelCase_ : Optional[int] = self.pipeline_class(**_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) UpperCAmelCase_ : Union[str, Any] = self.get_dummy_mask_inputs(_A ) UpperCAmelCase_ : int = pipe.generate_mask(**_A ) UpperCAmelCase_ : Tuple = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) UpperCAmelCase_ : List[Any] = np.array([0] * 9 ) UpperCAmelCase_ : Dict = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(_A , 1e-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def A ( self : str ) -> Optional[int]: UpperCAmelCase_ : Union[str, Any] = '''cpu''' UpperCAmelCase_ : str = self.get_dummy_components() UpperCAmelCase_ : str = self.pipeline_class(**_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) UpperCAmelCase_ : Optional[Any] = self.get_dummy_inversion_inputs(_A ) UpperCAmelCase_ : Optional[Any] = pipe.invert(**_A ).images UpperCAmelCase_ : List[Any] = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) UpperCAmelCase_ : int = np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) UpperCAmelCase_ : List[str] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_A , 1e-3 ) def A ( self : Tuple ) -> Optional[Any]: super().test_inference_batch_single_identical(expected_max_diff=5e-3 ) def A ( self : str ) -> Tuple: UpperCAmelCase_ : Any = '''cpu''' UpperCAmelCase_ : Union[str, Any] = self.get_dummy_components() UpperCAmelCase_ : Any = {'''beta_start''': 0.00_085, '''beta_end''': 0.012, '''beta_schedule''': '''scaled_linear'''} UpperCAmelCase_ : Any = DPMSolverMultistepScheduler(**_A ) UpperCAmelCase_ : Optional[Any] = DPMSolverMultistepInverseScheduler(**_A ) UpperCAmelCase_ : Union[str, Any] = self.pipeline_class(**_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) UpperCAmelCase_ : Union[str, Any] = self.get_dummy_inversion_inputs(_A ) UpperCAmelCase_ : Optional[Any] = pipe.invert(**_A ).images UpperCAmelCase_ : Tuple = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) UpperCAmelCase_ : List[Any] = np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) UpperCAmelCase_ : Optional[int] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_A , 1e-3 ) @require_torch_gpu @slow class snake_case__ ( unittest.TestCase): def A ( self : Optional[Any] ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def A ( cls : Dict ) -> List[Any]: UpperCAmelCase_ : Optional[int] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png''' ) UpperCAmelCase_ : int = raw_image.convert('''RGB''' ).resize((7_68, 7_68) ) UpperCAmelCase_ : Any = raw_image def A ( self : List[Any] ) -> List[str]: UpperCAmelCase_ : int = torch.manual_seed(0 ) UpperCAmelCase_ : str = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''' , safety_checker=_A , torch_dtype=torch.floataa ) UpperCAmelCase_ : List[str] = DDIMScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase_ : List[str] = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_A ) UpperCAmelCase_ : Optional[Any] = '''a bowl of fruit''' UpperCAmelCase_ : Tuple = '''a bowl of pears''' UpperCAmelCase_ : Optional[int] = pipe.generate_mask( image=self.raw_image , source_prompt=_A , target_prompt=_A , generator=_A , ) UpperCAmelCase_ : List[str] = pipe.invert( prompt=_A , image=self.raw_image , inpaint_strength=0.7 , generator=_A ).latents UpperCAmelCase_ : Any = pipe( prompt=_A , mask_image=_A , image_latents=_A , generator=_A , negative_prompt=_A , inpaint_strength=0.7 , output_type='''numpy''' , ).images[0] UpperCAmelCase_ : str = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((7_68, 7_68) ) ) / 2_55 ) assert np.abs((expected_image - image).max() ) < 5e-1 def A ( self : Tuple ) -> List[str]: UpperCAmelCase_ : Dict = torch.manual_seed(0 ) UpperCAmelCase_ : Any = StableDiffusionDiffEditPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-1''' , safety_checker=_A , torch_dtype=torch.floataa ) UpperCAmelCase_ : List[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase_ : Union[str, Any] = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=_A ) UpperCAmelCase_ : Optional[Any] = '''a bowl of fruit''' UpperCAmelCase_ : Dict = '''a bowl of pears''' UpperCAmelCase_ : Union[str, Any] = pipe.generate_mask( image=self.raw_image , source_prompt=_A , target_prompt=_A , generator=_A , ) UpperCAmelCase_ : List[Any] = pipe.invert( prompt=_A , image=self.raw_image , inpaint_strength=0.7 , generator=_A , num_inference_steps=25 , ).latents UpperCAmelCase_ : Dict = pipe( prompt=_A , mask_image=_A , image_latents=_A , generator=_A , negative_prompt=_A , inpaint_strength=0.7 , num_inference_steps=25 , output_type='''numpy''' , ).images[0] UpperCAmelCase_ : Tuple = ( np.array( load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/diffedit/pears.png''' ).resize((7_68, 7_68) ) ) / 2_55 ) assert np.abs((expected_image - image).max() ) < 5e-1
304
1
"""simple docstring""" import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format='%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s', datefmt='%Y-%m-%d %H:%M:%S', level=os.environ.get('LOGLEVEL', 'INFO').upper(), stream=sys.stdout, ) UpperCamelCase_ = logging.getLogger(__name__) UpperCamelCase_ = {'facebook/bart-base': BartForConditionalGeneration} UpperCamelCase_ = {'facebook/bart-base': BartTokenizer} def UpperCamelCase ( ) ->List[str]: """simple docstring""" a_ = argparse.ArgumentParser(description="Export Bart model + Beam Search to ONNX graph." ) parser.add_argument( "--validation_file" , type=UpperCAmelCase , default=UpperCAmelCase , help="A csv or a json file containing the validation data." ) parser.add_argument( "--max_length" , type=UpperCAmelCase , default=5 , help="The maximum total input sequence length after tokenization." , ) parser.add_argument( "--num_beams" , type=UpperCAmelCase , default=UpperCAmelCase , help=( "Number of beams to use for evaluation. This argument will be " "passed to ``model.generate``, which is used during ``evaluate`` and ``predict``." ) , ) parser.add_argument( "--model_name_or_path" , type=UpperCAmelCase , help="Path to pretrained model or model identifier from huggingface.co/models." , required=UpperCAmelCase , ) parser.add_argument( "--config_name" , type=UpperCAmelCase , default=UpperCAmelCase , help="Pretrained config name or path if not the same as model_name" , ) parser.add_argument( "--device" , type=UpperCAmelCase , default="cpu" , help="Device where the model will be run" , ) parser.add_argument("--output_file_path" , type=UpperCAmelCase , default=UpperCAmelCase , help="Where to store the final ONNX file." ) a_ = parser.parse_args() return args def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase="cpu" ) ->Optional[Any]: """simple docstring""" a_ = model_dict[model_name].from_pretrained(UpperCAmelCase ).to(UpperCAmelCase ) a_ = tokenizer_dict[model_name].from_pretrained(UpperCAmelCase ) if model_name in ["facebook/bart-base"]: a_ = 0 a_ = None a_ = 0 return huggingface_model, tokenizer def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->Dict: """simple docstring""" model.eval() a_ = None a_ = torch.jit.script(BARTBeamSearchGenerator(UpperCAmelCase ) ) with torch.no_grad(): a_ = "My friends are cool but they eat too many carbs." a_ = tokenizer([ARTICLE_TO_SUMMARIZE] , max_length=1_024 , return_tensors="pt" ).to(model.device ) a_ = model.generate( inputs["input_ids"] , attention_mask=inputs["attention_mask"] , num_beams=UpperCAmelCase , max_length=UpperCAmelCase , early_stopping=UpperCAmelCase , decoder_start_token_id=model.config.decoder_start_token_id , ) torch.onnx.export( UpperCAmelCase , ( inputs["input_ids"], inputs["attention_mask"], num_beams, max_length, model.config.decoder_start_token_id, ) , UpperCAmelCase , opset_version=14 , input_names=["input_ids", "attention_mask", "num_beams", "max_length", "decoder_start_token_id"] , output_names=["output_ids"] , dynamic_axes={ "input_ids": {0: "batch", 1: "seq"}, "output_ids": {0: "batch", 1: "seq_out"}, } , example_outputs=UpperCAmelCase , ) logger.info("Model exported to {}".format(UpperCAmelCase ) ) a_ = remove_dup_initializers(os.path.abspath(UpperCAmelCase ) ) logger.info("Deduplicated and optimized model written to {}".format(UpperCAmelCase ) ) a_ = onnxruntime.InferenceSession(UpperCAmelCase ) a_ = ort_sess.run( UpperCAmelCase , { "input_ids": inputs["input_ids"].cpu().numpy(), "attention_mask": inputs["attention_mask"].cpu().numpy(), "num_beams": np.array(UpperCAmelCase ), "max_length": np.array(UpperCAmelCase ), "decoder_start_token_id": np.array(model.config.decoder_start_token_id ), } , ) np.testing.assert_allclose(summary_ids.cpu().numpy() , ort_out[0] , rtol=1E-3 , atol=1E-3 ) logger.info("Model outputs from torch and ONNX Runtime are similar." ) logger.info("Success." ) def UpperCamelCase ( ) ->Any: """simple docstring""" a_ = parse_args() a_ = 5 a_ = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.setLevel(logging.INFO ) transformers.utils.logging.set_verbosity_error() a_ = torch.device(args.device ) a_ , a_ = load_model_tokenizer(args.model_name_or_path , UpperCAmelCase ) if model.config.decoder_start_token_id is None: raise ValueError("Make sure that `config.decoder_start_token_id` is correctly defined" ) model.to(UpperCAmelCase ) if args.max_length: a_ = args.max_length if args.num_beams: a_ = args.num_beams if args.output_file_path: a_ = args.output_file_path else: a_ = "BART.onnx" logger.info("Exporting model to ONNX" ) export_and_validate_model(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if __name__ == "__main__": main()
303
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase ( UpperCAmelCase ) ->Tuple: """simple docstring""" a_ = [2, 2, 6, 2] if "tiny" in model_name else [2, 2, 18, 2] a_ = True if "large" in model_name or "huge" in model_name else False a_ = True if "large" in model_name or "huge" in model_name else False a_ = True if "large" in model_name or "huge" in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: a_ = [3, 3, 3, 3] a_ = [5, 5, 5, 5] elif "fl4" in model_name: a_ = [4, 4, 4, 4] a_ = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: a_ = [3, 3, 3, 3] if "lrf" in model_name: a_ = [3, 3, 3, 3] else: a_ = [2, 2, 2, 2] if "tiny" in model_name: a_ = 96 elif "small" in model_name: a_ = 96 elif "base" in model_name: a_ = 128 elif "large" in model_name: a_ = 192 elif "xlarge" in model_name: a_ = 256 elif "huge" in model_name: a_ = 352 # set label information a_ = "huggingface/label-files" if "large" in model_name or "huge" in model_name: a_ = "imagenet-22k-id2label.json" else: a_ = "imagenet-1k-id2label.json" a_ = json.load(open(hf_hub_download(UpperCAmelCase , UpperCAmelCase , repo_type="dataset" ) , "r" ) ) a_ = {int(UpperCAmelCase ): v for k, v in idalabel.items()} a_ = {v: k for k, v in idalabel.items()} a_ = FocalNetConfig( embed_dim=UpperCAmelCase , depths=UpperCAmelCase , focal_levels=UpperCAmelCase , focal_windows=UpperCAmelCase , use_conv_embed=UpperCAmelCase , idalabel=UpperCAmelCase , labelaid=UpperCAmelCase , use_post_layernorm=UpperCAmelCase , use_layerscale=UpperCAmelCase , ) return config def UpperCamelCase ( UpperCAmelCase ) ->Any: """simple docstring""" if "patch_embed.proj" in name: a_ = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: a_ = name.replace("patch_embed.norm" , "embeddings.norm" ) if "layers" in name: a_ = "encoder." + name if "encoder.layers" in name: a_ = name.replace("encoder.layers" , "encoder.stages" ) if "downsample.proj" in name: a_ = name.replace("downsample.proj" , "downsample.projection" ) if "blocks" in name: a_ = name.replace("blocks" , "layers" ) if "modulation.f.weight" in name or "modulation.f.bias" in name: a_ = name.replace("modulation.f" , "modulation.projection_in" ) if "modulation.h.weight" in name or "modulation.h.bias" in name: a_ = name.replace("modulation.h" , "modulation.projection_context" ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: a_ = name.replace("modulation.proj" , "modulation.projection_out" ) if name == "norm.weight": a_ = "layernorm.weight" if name == "norm.bias": a_ = "layernorm.bias" if "head" in name: a_ = name.replace("head" , "classifier" ) else: a_ = "focalnet." + name return name def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=False ) ->Dict: """simple docstring""" a_ = { "focalnet-tiny": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth", "focalnet-tiny-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth", "focalnet-small": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth", "focalnet-small-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth", "focalnet-base": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth", "focalnet-base-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth", "focalnet-large-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth", "focalnet-large-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth", "focalnet-xlarge-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth", "focalnet-xlarge-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth", } # fmt: on a_ = model_name_to_url[model_name] print("Checkpoint URL: " , UpperCAmelCase ) a_ = torch.hub.load_state_dict_from_url(UpperCAmelCase , map_location="cpu" )["model"] # rename keys for key in state_dict.copy().keys(): a_ = state_dict.pop(UpperCAmelCase ) a_ = val a_ = get_focalnet_config(UpperCAmelCase ) a_ = FocalNetForImageClassification(UpperCAmelCase ) model.eval() # load state dict model.load_state_dict(UpperCAmelCase ) # verify conversion a_ = "http://images.cocodataset.org/val2017/000000039769.jpg" a_ = BitImageProcessor( do_resize=UpperCAmelCase , size={"shortest_edge": 256} , resample=PILImageResampling.BILINEAR , do_center_crop=UpperCAmelCase , crop_size=224 , do_normalize=UpperCAmelCase , image_mean=UpperCAmelCase , image_std=UpperCAmelCase , ) a_ = Image.open(requests.get(UpperCAmelCase , stream=UpperCAmelCase ).raw ) a_ = processor(images=UpperCAmelCase , return_tensors="pt" ) a_ = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) a_ = image_transforms(UpperCAmelCase ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , UpperCAmelCase , atol=1E-4 ) a_ = model(**UpperCAmelCase ) a_ = outputs.logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) print("First values of logits:" , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": a_ = torch.tensor([0.2166, -0.4368, 0.2191] ) elif model_name == "focalnet-tiny-lrf": a_ = torch.tensor([1.1669, 0.0125, -0.1695] ) elif model_name == "focalnet-small": a_ = torch.tensor([0.4917, -0.0430, 0.1341] ) elif model_name == "focalnet-small-lrf": a_ = torch.tensor([-0.2588, -0.5342, -0.2331] ) elif model_name == "focalnet-base": a_ = torch.tensor([-0.1655, -0.4090, -0.1730] ) elif model_name == "focalnet-base-lrf": a_ = torch.tensor([0.5306, -0.0483, -0.3928] ) assert torch.allclose(outputs.logits[0, :3] , UpperCAmelCase , atol=1E-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCAmelCase ) processor.save_pretrained(UpperCAmelCase ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='focalnet-tiny', type=str, help='Name of the FocalNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub.', ) UpperCamelCase_ = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
303
1
import os import re import shutil import sys import tempfile import unittest import black A_ : List[str] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. A_ : str = ' def __init__(self, config):\n super().__init__()\n self.transform = BertPredictionHeadTransform(config)\n\n # The output weights are the same as the input embeddings, but there is\n # an output-only bias for each token.\n self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False)\n\n self.bias = nn.Parameter(torch.zeros(config.vocab_size))\n\n # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings`\n self.decoder.bias = self.bias\n\n def forward(self, hidden_states):\n hidden_states = self.transform(hidden_states)\n hidden_states = self.decoder(hidden_states)\n return hidden_states\n' class _a (unittest.TestCase ): '''simple docstring''' def __A ( self ): A__ : int = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , """models/bert/""" ) ) A__ : Dict = self.transformer_dir shutil.copy( os.path.join(A__ , """src/transformers/models/bert/modeling_bert.py""" ) , os.path.join(self.transformer_dir , """models/bert/modeling_bert.py""" ) , ) def __A ( self ): A__ : Any = """src/transformers""" shutil.rmtree(self.transformer_dir ) def __A ( self , A__ , A__ , A__ , A__=None ): A__ : Optional[int] = comment + F"""\nclass {class_name}(nn.Module):\n""" + class_code if overwrite_result is not None: A__ : List[str] = comment + F"""\nclass {class_name}(nn.Module):\n""" + overwrite_result A__ : Optional[int] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) A__ : Union[str, Any] = black.format_str(A__ , mode=A__ ) A__ : Union[str, Any] = os.path.join(self.transformer_dir , """new_code.py""" ) with open(A__ , """w""" , newline="""\n""" ) as f: f.write(A__ ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(A__ ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=A__ ) with open(A__ , """r""" ) as f: self.assertTrue(f.read() , A__ ) def __A ( self ): A__ : str = check_copies.find_code_in_transformers("""models.bert.modeling_bert.BertLMPredictionHead""" ) self.assertEqual(A__ , A__ ) def __A ( self ): # Base copy consistency self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead""" , """BertLMPredictionHead""" , REFERENCE_CODE + """\n""" , ) # With no empty line at the end self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead""" , """BertLMPredictionHead""" , A__ , ) # Copy consistency with rename self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel""" , """TestModelLMPredictionHead""" , re.sub("""Bert""" , """TestModel""" , A__ ) , ) # Copy consistency with a really long name A__ : Tuple = """TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason""" self.check_copy_consistency( F"""# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}""" , F"""{long_class_name}LMPredictionHead""" , re.sub("""Bert""" , A__ , A__ ) , ) # Copy consistency with overwrite self.check_copy_consistency( """# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel""" , """TestModelLMPredictionHead""" , A__ , overwrite_result=re.sub("""Bert""" , """TestModel""" , A__ ) , ) def __A ( self ): A__ : Union[str, Any] = check_copies.LOCALIZED_READMES["""README_zh-hans.md"""] A__ : Dict = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the""" """ Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for""" """ Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong""" """ Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.""" """ **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),""" """ released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and""" """ lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same""" """ method has been applied to compress GPT2 into""" """ [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into""" """ [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),""" """ Multilingual BERT into""" """ [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German""" """ version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**""" """ (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders""" """ as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang""" """ Luong, Quoc V. Le, Christopher D. Manning.""" ) A__ : str = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ : Union[str, Any] = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.""" """ **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文""" """ [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and""" """ lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same""" """ method has been applied to compress GPT2 into""" """ [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into""" """ [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),""" """ Multilingual BERT into""" """ [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German""" """ version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自""" """ Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather""" """ than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,""" """ Christopher D. Manning 发布。\n""" ) A__ , A__ : List[Any] = check_copies.convert_to_localized_md( A__ , A__ , localized_readme["""format_model_list"""] ) self.assertFalse(A__ ) self.assertEqual(A__ , A__ ) A__ , A__ : List[Any] = check_copies.convert_to_localized_md( A__ , A__ , localized_readme["""format_model_list"""] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(A__ ) A__ : List[str] = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the""" """ Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for""" """ Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong""" """ Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.""" ) A__ : Optional[int] = ( """1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and""" """ the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ : Union[str, Any] = ( """1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the""" """ Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of""" """ Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian""" """ Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n""" ) A__ , A__ : Optional[int] = check_copies.convert_to_localized_md( A__ , A__ , localized_readme["""format_model_list"""] ) # Check if the model link is synchronized. self.assertEqual(A__ , A__ )
192
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def UpperCamelCase (lowercase_: int , lowercase_: Dict , lowercase_: Tuple ) -> Any: # Construct model if gpta_config_file == "": A__ : Dict = GPTaConfig() else: A__ : List[Any] = GPTaConfig.from_json_file(lowercase_ ) A__ : Tuple = GPTaModel(lowercase_ ) # Load weights from numpy load_tf_weights_in_gpta(lowercase_ , lowercase_ , lowercase_ ) # Save pytorch-model A__ : Optional[Any] = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME A__ : Optional[Any] = pytorch_dump_folder_path + """/""" + CONFIG_NAME print(f"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(model.state_dict() , lowercase_ ) print(f"""Save configuration file to {pytorch_config_dump_path}""" ) with open(lowercase_ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A_ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--gpt2_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--gpt2_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained OpenAI model. \n' 'This specifies the model architecture.' ), ) A_ : str = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
192
1
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: _UpperCAmelCase : Optional[int] = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class __lowerCAmelCase ( unittest.TestCase): def __init__( self: Optional[int] , _lowerCAmelCase: Union[str, Any] , _lowerCAmelCase: Optional[Any]=7 , _lowerCAmelCase: Optional[int]=3 , _lowerCAmelCase: List[Any]=18 , _lowerCAmelCase: Tuple=30 , _lowerCAmelCase: Tuple=4_00 , _lowerCAmelCase: Dict=None , _lowerCAmelCase: List[Any]=True , _lowerCAmelCase: Dict=True , _lowerCAmelCase: Optional[Any]=None , ): lowercase :List[str] = size if size is not None else {'height': 20, 'width': 20} lowercase :Dict = parent lowercase :Union[str, Any] = batch_size lowercase :Optional[Any] = num_channels lowercase :Union[str, Any] = image_size lowercase :Tuple = min_resolution lowercase :Optional[Any] = max_resolution lowercase :Union[str, Any] = size lowercase :Dict = do_normalize lowercase :Dict = do_convert_rgb lowercase :Union[str, Any] = [5_12, 10_24, 20_48, 40_96] lowercase :Optional[int] = patch_size if patch_size is not None else {'height': 16, 'width': 16} def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def SCREAMING_SNAKE_CASE ( self: Any ): lowercase :str = 'https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg' lowercase :Optional[Any] = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ).convert("RGB" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class __lowerCAmelCase ( lowerCAmelCase , unittest.TestCase): _a = PixaStructImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self: Tuple ): lowercase :Optional[Any] = PixaStructImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self: Tuple ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): lowercase :Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , "do_normalize" ) ) self.assertTrue(hasattr(_lowerCAmelCase , "do_convert_rgb" ) ) def SCREAMING_SNAKE_CASE ( self: str ): lowercase :Union[str, Any] = self.image_processor_tester.prepare_dummy_image() lowercase :Any = self.image_processing_class(**self.image_processor_dict ) lowercase :List[Any] = 20_48 lowercase :str = image_processor(_lowerCAmelCase , return_tensors="pt" , max_patches=_lowerCAmelCase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.06_06 ) , atol=1e-3 , rtol=1e-3 ) ) def SCREAMING_SNAKE_CASE ( self: int ): # Initialize image_processor lowercase :str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase :Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , Image.Image ) # Test not batched input lowercase :int = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase :Optional[int] = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched lowercase :Union[str, Any] = image_processor( _lowerCAmelCase , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): # Initialize image_processor lowercase :Any = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase :Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , Image.Image ) # Test not batched input lowercase :Optional[int] = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 lowercase :List[Any] = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(_lowerCAmelCase ): lowercase :Union[str, Any] = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches lowercase :Optional[int] = 'Hello' lowercase :int = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_lowerCAmelCase , header_text=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched lowercase :List[Any] = image_processor( _lowerCAmelCase , return_tensors="pt" , max_patches=_lowerCAmelCase , header_text=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def SCREAMING_SNAKE_CASE ( self: Optional[int] ): # Initialize image_processor lowercase :str = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowercase :str = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , numpify=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , np.ndarray ) lowercase :Tuple = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase :Dict = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched lowercase :Dict = image_processor( _lowerCAmelCase , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def SCREAMING_SNAKE_CASE ( self: Dict ): # Initialize image_processor lowercase :Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowercase :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , torchify=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , torch.Tensor ) # Test not batched input lowercase :Tuple = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase :Optional[Any] = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched lowercase :Dict = image_processor( _lowerCAmelCase , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class __lowerCAmelCase ( lowerCAmelCase , unittest.TestCase): _a = PixaStructImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self: List[str] ): lowercase :str = PixaStructImageProcessingTester(self , num_channels=4 ) lowercase :Union[str, Any] = 3 @property def SCREAMING_SNAKE_CASE ( self: str ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self: List[Any] ): lowercase :Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , "do_normalize" ) ) self.assertTrue(hasattr(_lowerCAmelCase , "do_convert_rgb" ) ) def SCREAMING_SNAKE_CASE ( self: str ): # Initialize image_processor lowercase :Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowercase :List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , Image.Image ) # Test not batched input lowercase :Dict = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowercase :int = image_processor( image_inputs[0] , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched lowercase :int = image_processor( _lowerCAmelCase , return_tensors="pt" , max_patches=_lowerCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
362
import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): # Construct model if openai_config_file == "": lowercase :List[Any] = OpenAIGPTConfig() else: lowercase :int = OpenAIGPTConfig.from_json_file(lowerCamelCase ) lowercase :Union[str, Any] = OpenAIGPTModel(lowerCamelCase ) # Load weights from numpy load_tf_weights_in_openai_gpt(lowerCamelCase, lowerCamelCase, lowerCamelCase ) # Save pytorch-model lowercase :Optional[Any] = pytorch_dump_folder_path + "/" + WEIGHTS_NAME lowercase :Any = pytorch_dump_folder_path + "/" + CONFIG_NAME print(F"Save PyTorch model to {pytorch_weights_dump_path}" ) torch.save(model.state_dict(), lowerCamelCase ) print(F"Save configuration file to {pytorch_config_dump_path}" ) with open(lowerCamelCase, "w", encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--openai_checkpoint_folder_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--openai_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained OpenAI model. \n" "This specifies the model architecture." ), ) _UpperCAmelCase : List[Any] = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
158
0
import pytest _snake_case = "__dummy_dataset1__" _snake_case = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn-validation.jsonl\"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\n \"ner_tags\": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n \"O\",\n \"B-PER\",\n \"I-PER\",\n \"B-ORG\",\n \"I-ORG\",\n \"B-LOC\",\n \"I-LOC\",\n ]\n )\n ),\n \"langs\": datasets.Sequence(datasets.Value(\"string\")),\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={\"filepath\": dl_path[\"train\"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={\"filepath\": dl_path[\"validation\"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n" @pytest.fixture def A ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def A ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = dataset_loading_script_name _lowerCAmelCase : Optional[int] = tmp_path / '''datasets''' / script_name script_dir.mkdir(parents=_lowerCamelCase ) _lowerCAmelCase : Dict = script_dir / F"{script_name}.py" with open(_lowerCamelCase , "w" ) as f: f.write(_lowerCamelCase ) return str(_lowerCamelCase )
36
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __magic_name__ ( unittest.TestCase): def UpperCAmelCase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Optional[int] = '''ZinengTang/tvlt-base''' UpperCamelCase__ : int = tempfile.mkdtemp() def UpperCAmelCase__ ( self : int , **lowerCamelCase__ : List[str] ) -> List[Any]: '''simple docstring''' return TvltImageProcessor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : Optional[Any] , **lowerCamelCase__ : Tuple ) -> List[Any]: '''simple docstring''' return TvltFeatureExtractor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : str ) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' UpperCamelCase__ : int = self.get_image_processor() UpperCamelCase__ : Union[str, Any] = self.get_feature_extractor() UpperCamelCase__ : List[str] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase__ : Optional[int] = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , lowerCamelCase__ ) self.assertIsInstance(processor.image_processor , lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] ) -> Tuple: '''simple docstring''' UpperCamelCase__ : str = self.get_image_processor() UpperCamelCase__ : List[Any] = self.get_feature_extractor() UpperCamelCase__ : Dict = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : Any = np.ones([12000] ) UpperCamelCase__ : Union[str, Any] = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : Any = processor(audio=lowerCamelCase__ , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : List[Any] = self.get_image_processor() UpperCamelCase__ : Any = self.get_feature_extractor() UpperCamelCase__ : int = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : int = np.ones([3, 224, 224] ) UpperCamelCase__ : List[str] = image_processor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : str = processor(images=lowerCamelCase__ , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Any = self.get_image_processor() UpperCamelCase__ : Dict = self.get_feature_extractor() UpperCamelCase__ : Union[str, Any] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : List[str] = np.ones([12000] ) UpperCamelCase__ : Tuple = np.ones([3, 224, 224] ) UpperCamelCase__ : Optional[Any] = processor(audio=lowerCamelCase__ , images=lowerCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(lowerCamelCase__ ): processor() def UpperCAmelCase__ ( self : Dict ) -> int: '''simple docstring''' UpperCamelCase__ : List[str] = self.get_image_processor() UpperCamelCase__ : str = self.get_feature_extractor() UpperCamelCase__ : Tuple = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
146
0
'''simple docstring''' from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ) -> List[str]: '''simple docstring''' return {key.lstrip('''-''' ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def _UpperCamelCase ( ) -> str: '''simple docstring''' snake_case : str = ArgumentParser( '''HuggingFace Datasets CLI tool''' , usage='''datasets-cli <command> [<args>]''' , allow_abbrev=SCREAMING_SNAKE_CASE__ ) snake_case : Optional[Any] = parser.add_subparsers(help='''datasets-cli command helpers''' ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) EnvironmentCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) TestCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) RunBeamCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) DummyDataCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) # Parse args snake_case : Optional[int] = parser.parse_known_args() if not hasattr(SCREAMING_SNAKE_CASE__ , '''func''' ): parser.print_help() exit(1 ) snake_case : Dict = parse_unknown_args(SCREAMING_SNAKE_CASE__ ) # Run snake_case : Optional[Any] = args.func(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) service.run() if __name__ == "__main__": main()
369
'''simple docstring''' from functools import lru_cache @lru_cache def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ) -> int: '''simple docstring''' if num < 0: raise ValueError('''Number should not be negative.''' ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
83
0
"""simple docstring""" from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration __UpperCamelCase : str = HfArgumentParser(InitializationArguments) __UpperCamelCase : List[str] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization __UpperCamelCase : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks __UpperCamelCase : Optional[int] = { '''vocab_size''': len(tokenizer), '''scale_attn_by_inverse_layer_idx''': True, '''reorder_and_upcast_attn''': True, } # Load model config (GPT-2 large in this case) __UpperCamelCase : Tuple = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config __UpperCamelCase : Optional[int] = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
106
"""simple docstring""" import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=99 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=50 , lowerCamelCase=0.02 , lowerCamelCase=True , lowerCamelCase=None , ): __a = parent __a = batch_size __a = seq_length __a = is_training __a = use_input_mask __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = initializer_range __a = use_labels __a = scope def a__ ( self ): __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = None if self.use_input_mask: __a = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = self.get_config() return config, input_ids, input_mask, token_labels def a__ ( self ): return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , ) def a__ ( self ): ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = self.prepare_config_and_inputs() __a = True __a = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __a = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): __a = BertGenerationEncoder(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , attention_mask=lowerCamelCase ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): __a = True __a = BertGenerationEncoder(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , ) __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): __a = True __a = True __a = BertGenerationDecoder(config=lowerCamelCase ).to(lowerCamelCase ).eval() # first forward pass __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , use_cache=lowerCamelCase , ) __a = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __a = ids_tensor((self.batch_size, 3) , config.vocab_size ) __a = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __a = torch.cat([input_ids, next_tokens] , dim=-1 ) __a = torch.cat([input_mask, next_mask] , dim=-1 ) __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] # select random slice __a = ids_tensor((1,) , output_from_past.shape[-1] ).item() __a = output_from_no_past[:, -3:, random_slice_idx].detach() __a = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , *lowerCamelCase , ): __a = BertGenerationDecoder(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__ ( self ): __a , __a , __a , __a = self.prepare_config_and_inputs() __a = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : Union[str, Any] = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () _snake_case : Any = (BertGenerationDecoder,) if is_torch_available() else () _snake_case : Union[str, Any] = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def a__ ( self ): __a = BertGenerationEncoderTester(self ) __a = ConfigTester(self , config_class=lowerCamelCase , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def a__ ( self ): __a , __a , __a , __a = self.model_tester.prepare_config_and_inputs() __a = "bert" self.model_tester.create_and_check_model(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*lowerCamelCase ) def a__ ( self ): # This regression test was failing with PyTorch < 1.3 ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() __a = None self.model_tester.create_and_check_model_as_decoder( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*lowerCamelCase ) @slow def a__ ( self ): __a = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) self.assertIsNotNone(lowerCamelCase ) @require_torch class snake_case__ ( unittest.TestCase ): @slow def a__ ( self ): __a = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) __a = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 10140, 102]] ) with torch.no_grad(): __a = model(lowerCamelCase )[0] __a = torch.Size([1, 8, 1024] ) self.assertEqual(output.shape , lowerCamelCase ) __a = torch.tensor( [[[0.1775, 0.0083, -0.0321], [1.6002, 0.1287, 0.3912], [2.1473, 0.5791, 0.6066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @require_torch class snake_case__ ( unittest.TestCase ): @slow def a__ ( self ): __a = BertGenerationDecoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) __a = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 10140, 102]] ) with torch.no_grad(): __a = model(lowerCamelCase )[0] __a = torch.Size([1, 8, 50358] ) self.assertEqual(output.shape , lowerCamelCase ) __a = torch.tensor( [[[-0.5788, -2.5994, -3.7054], [0.0438, 4.7997, 1.8795], [1.5862, 6.6409, 4.4638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase , atol=1E-4 ) )
261
0
"""simple docstring""" import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class a ( lowerCAmelCase_ , unittest.TestCase ): # TODO: is there an appropriate internal test set? _snake_case : Tuple = 'ssube/stable-diffusion-x4-upscaler-onnx' def lowerCAmelCase_ ( self : Tuple , __lowerCAmelCase : Any=0 ): _UpperCAmelCase = floats_tensor((1, 3, 128, 128) , rng=random.Random(__lowerCAmelCase ) ) _UpperCAmelCase = torch.manual_seed(__lowerCAmelCase ) _UpperCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def lowerCAmelCase_ ( self : List[str] ): _UpperCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _UpperCAmelCase = self.get_dummy_inputs() _UpperCAmelCase = pipe(**__lowerCAmelCase ).images _UpperCAmelCase = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array( [0.6_974_782, 0.68_902_093, 0.70_135_885, 0.7_583_618, 0.7_804_545, 0.7_854_912, 0.78_667_426, 0.78_743_863, 0.78_070_223] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def lowerCAmelCase_ ( self : int ): _UpperCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _UpperCAmelCase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _UpperCAmelCase = self.get_dummy_inputs() _UpperCAmelCase = pipe(**__lowerCAmelCase ).images _UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array( [0.6_898_892, 0.59_240_556, 0.52_499_527, 0.58_866_215, 0.52_258_235, 0.52_572_715, 0.62_414_473, 0.6_174_387, 0.6_214_964] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def lowerCAmelCase_ ( self : Union[str, Any] ): _UpperCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _UpperCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _UpperCAmelCase = self.get_dummy_inputs() _UpperCAmelCase = pipe(**__lowerCAmelCase ).images _UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array( [0.7_659_278, 0.76_437_664, 0.75_579_107, 0.7_691_116, 0.77_666_986, 0.7_727_672, 0.7_758_664, 0.7_812_226, 0.76_942_515] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def lowerCAmelCase_ ( self : Dict ): _UpperCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _UpperCAmelCase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _UpperCAmelCase = self.get_dummy_inputs() _UpperCAmelCase = pipe(**__lowerCAmelCase ).images _UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array( [0.6_974_782, 0.68_902_093, 0.70_135_885, 0.7_583_618, 0.7_804_545, 0.7_854_912, 0.78_667_426, 0.78_743_863, 0.78_070_223] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def lowerCAmelCase_ ( self : List[str] ): _UpperCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _UpperCAmelCase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _UpperCAmelCase = self.get_dummy_inputs() _UpperCAmelCase = pipe(**__lowerCAmelCase ).images _UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array( [0.77_424_496, 0.773_601, 0.7_645_288, 0.7_769_598, 0.7_772_739, 0.7_738_688, 0.78_187_233, 0.77_879_584, 0.767_043] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class a ( unittest.TestCase ): @property def lowerCAmelCase_ ( self : str ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCAmelCase_ ( self : str ): _UpperCAmelCase = ort.SessionOptions() _UpperCAmelCase = False return options def lowerCAmelCase_ ( self : Optional[Any] ): _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) _UpperCAmelCase = init_image.resize((128, 128) ) # using the PNDM scheduler by default _UpperCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained( """ssube/stable-diffusion-x4-upscaler-onnx""" , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _UpperCAmelCase = """A fantasy landscape, trending on artstation""" _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = pipe( prompt=__lowerCAmelCase , image=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=10 , generator=__lowerCAmelCase , output_type="""np""" , ) _UpperCAmelCase = output.images _UpperCAmelCase = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array([0.4_883, 0.4_947, 0.4_980, 0.4_975, 0.4_982, 0.4_980, 0.5_000, 0.5_006, 0.4_972] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def lowerCAmelCase_ ( self : Union[str, Any] ): _UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) _UpperCAmelCase = init_image.resize((128, 128) ) _UpperCAmelCase = LMSDiscreteScheduler.from_pretrained( """ssube/stable-diffusion-x4-upscaler-onnx""" , subfolder="""scheduler""" ) _UpperCAmelCase = OnnxStableDiffusionUpscalePipeline.from_pretrained( """ssube/stable-diffusion-x4-upscaler-onnx""" , scheduler=__lowerCAmelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _UpperCAmelCase = """A fantasy landscape, trending on artstation""" _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = pipe( prompt=__lowerCAmelCase , image=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=20 , generator=__lowerCAmelCase , output_type="""np""" , ) _UpperCAmelCase = output.images _UpperCAmelCase = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array( [0.50_173_753, 0.50_223_356, 0.502_039, 0.50_233_036, 0.5_023_725, 0.5_022_601, 0.5_018_758, 0.50_234_085, 0.50_241_566] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
30
"""simple docstring""" import mpmath # for roots of unity import numpy as np class a : def __init__( self : Tuple , __lowerCAmelCase : Dict=None , __lowerCAmelCase : Union[str, Any]=None ): # Input as list _UpperCAmelCase = list(poly_a or [0] )[:] _UpperCAmelCase = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() _UpperCAmelCase = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() _UpperCAmelCase = len(self.polyB ) # Add 0 to make lengths equal a power of 2 _UpperCAmelCase = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform _UpperCAmelCase = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product _UpperCAmelCase = self.__multiply() def lowerCAmelCase_ ( self : Optional[int] , __lowerCAmelCase : str ): _UpperCAmelCase = [[x] for x in self.polyA] if which == """A""" else [[x] for x in self.polyB] # Corner case if len(__lowerCAmelCase ) <= 1: return dft[0] # _UpperCAmelCase = self.c_max_length // 2 while next_ncol > 0: _UpperCAmelCase = [[] for i in range(__lowerCAmelCase )] _UpperCAmelCase = self.root**next_ncol # First half of next step _UpperCAmelCase = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__lowerCAmelCase ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step _UpperCAmelCase = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__lowerCAmelCase ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update _UpperCAmelCase = new_dft _UpperCAmelCase = next_ncol // 2 return dft[0] def lowerCAmelCase_ ( self : Tuple ): _UpperCAmelCase = self.__dft("""A""" ) _UpperCAmelCase = self.__dft("""B""" ) _UpperCAmelCase = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT _UpperCAmelCase = 2 while next_ncol <= self.c_max_length: _UpperCAmelCase = [[] for i in range(__lowerCAmelCase )] _UpperCAmelCase = self.root ** (next_ncol // 2) _UpperCAmelCase = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update _UpperCAmelCase = new_inverse_c next_ncol *= 2 # Unpack _UpperCAmelCase = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Dict ): _UpperCAmelCase = """A = """ + """ + """.join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) _UpperCAmelCase = """B = """ + """ + """.join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) _UpperCAmelCase = """A*B = """ + """ + """.join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
30
1
import os def _snake_case( ) -> str: '''simple docstring''' with open(os.path.dirname(_a ) + '/p022_names.txt' ) as file: A__ = str(file.readlines()[0] ) A__ = names.replace('\"' , '' ).split(',' ) names.sort() A__ = 0 A__ = 0 for i, name in enumerate(_a ): for letter in name: name_score += ord(_a ) - 64 total_score += (i + 1) * name_score A__ = 0 return total_score if __name__ == "__main__": print(solution())
7
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =CustomTokenizer pass
76
0
import numpy as np class lowerCamelCase_ : '''simple docstring''' def __init__( self) -> List[Any]: __UpperCamelCase :List[str] = (0, 0) __UpperCamelCase :int = None __UpperCamelCase :Optional[int] = 0 __UpperCamelCase :Dict = 0 __UpperCamelCase :Optional[Any] = 0 def __eq__( self , __lowercase) -> Optional[Any]: return self.position == cell.position def UpperCamelCase__ ( self) -> Any: print(self.position) class lowerCamelCase_ : '''simple docstring''' def __init__( self , __lowercase=(5, 5)) -> Optional[int]: __UpperCamelCase :Dict = np.zeros(__lowercase) __UpperCamelCase :Tuple = world_size[0] __UpperCamelCase :List[Any] = world_size[1] def UpperCamelCase__ ( self) -> List[Any]: print(self.w) def UpperCamelCase__ ( self , __lowercase) -> Optional[Any]: __UpperCamelCase :Optional[int] = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] __UpperCamelCase :List[str] = cell.position[0] __UpperCamelCase :Dict = cell.position[1] __UpperCamelCase :int = [] for n in neughbour_cord: __UpperCamelCase :int = current_x + n[0] __UpperCamelCase :Tuple = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: __UpperCamelCase :int = Cell() __UpperCamelCase :Optional[int] = (x, y) __UpperCamelCase :Union[str, Any] = cell neighbours.append(__lowercase) return neighbours def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = [] __UpperCamelCase :str = [] _open.append(SCREAMING_SNAKE_CASE ) while _open: __UpperCamelCase :int = np.argmin([n.f for n in _open] ) __UpperCamelCase :Dict = _open[min_f] _closed.append(_open.pop(SCREAMING_SNAKE_CASE ) ) if current == goal: break for n in world.get_neigbours(SCREAMING_SNAKE_CASE ): for c in _closed: if c == n: continue __UpperCamelCase :List[Any] = current.g + 1 __UpperCamelCase :int = n.position __UpperCamelCase :List[Any] = goal.position __UpperCamelCase :List[str] = (ya - ya) ** 2 + (xa - xa) ** 2 __UpperCamelCase :List[str] = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = [] while current.parent is not None: path.append(current.position ) __UpperCamelCase :Optional[int] = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": __lowercase = Gridworld() # Start position and goal __lowercase = Cell() __lowercase = (0, 0) __lowercase = Cell() __lowercase = (4, 4) print(F'path from {start.position} to {goal.position}') __lowercase = astar(world, start, goal) # Just for visual reasons. for i in s: __lowercase = 1 print(world.w)
350
import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset __lowercase = '''bert-base-cased''' __lowercase = '''google/pegasus-xsum''' __lowercase = [''' Sam ate lunch today.''', '''Sams lunch ingredients.'''] __lowercase = ['''A very interesting story about what I ate for lunch.''', '''Avocado, celery, turkey, coffee'''] __lowercase = '''patrickvonplaten/t5-tiny-random''' __lowercase = '''sshleifer/bart-tiny-random''' __lowercase = '''sshleifer/tiny-mbart''' __lowercase = '''sshleifer/tiny-marian-en-de''' def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = '''\n'''.join(SCREAMING_SNAKE_CASE ) Path(SCREAMING_SNAKE_CASE ).open('''w''' ).writelines(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' for split in ["train", "val", "test"]: _dump_articles(os.path.join(SCREAMING_SNAKE_CASE , f"""{split}.source""" ) , SCREAMING_SNAKE_CASE ) _dump_articles(os.path.join(SCREAMING_SNAKE_CASE , f"""{split}.target""" ) , SCREAMING_SNAKE_CASE ) return tmp_dir class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) @slow def UpperCamelCase__ ( self , __lowercase) -> List[Any]: __UpperCamelCase :Dict = AutoTokenizer.from_pretrained(__lowercase) __UpperCamelCase :Optional[Any] = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir()) __UpperCamelCase :List[Any] = max(len(tokenizer.encode(__lowercase)) for a in ARTICLES) __UpperCamelCase :Optional[int] = max(len(tokenizer.encode(__lowercase)) for a in SUMMARIES) __UpperCamelCase :int = 4 __UpperCamelCase :Any = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated __UpperCamelCase , __UpperCamelCase :Tuple = '''ro_RO''', '''de_DE''' # ignored for all but mbart, but never causes error. __UpperCamelCase :str = SeqaSeqDataset( __lowercase , data_dir=__lowercase , type_path='''train''' , max_source_length=__lowercase , max_target_length=__lowercase , src_lang=__lowercase , tgt_lang=__lowercase , ) __UpperCamelCase :Any = DataLoader(__lowercase , batch_size=2 , collate_fn=train_dataset.collate_fn) for batch in dataloader: assert isinstance(__lowercase , __lowercase) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place __UpperCamelCase :Optional[int] = shift_tokens_right(batch['''labels'''] , tokenizer.pad_token_id) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED]) def UpperCamelCase__ ( self , __lowercase) -> int: __UpperCamelCase :Union[str, Any] = AutoTokenizer.from_pretrained(__lowercase) __UpperCamelCase :Union[str, Any] = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir()) __UpperCamelCase :int = max(len(tokenizer.encode(__lowercase)) for a in ARTICLES) __UpperCamelCase :Dict = max(len(tokenizer.encode(__lowercase)) for a in SUMMARIES) __UpperCamelCase :Union[str, Any] = 4 __UpperCamelCase :List[str] = LegacySeqaSeqDataset( __lowercase , data_dir=__lowercase , type_path='''train''' , max_source_length=20 , max_target_length=__lowercase , ) __UpperCamelCase :Dict = DataLoader(__lowercase , batch_size=2 , collate_fn=train_dataset.collate_fn) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 20 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :List[Any] = AutoTokenizer.from_pretrained('''facebook/mbart-large-cc25''') __UpperCamelCase :Union[str, Any] = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir())) __UpperCamelCase :str = tmp_dir.joinpath('''train.source''').open().readlines() __UpperCamelCase :int = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir())) pack_data_dir(__lowercase , __lowercase , 128 , __lowercase) __UpperCamelCase :Union[str, Any] = {x.name for x in tmp_dir.iterdir()} __UpperCamelCase :int = {x.name for x in save_dir.iterdir()} __UpperCamelCase :Optional[int] = save_dir.joinpath('''train.source''').open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(__lowercase) < len(__lowercase) assert len(__lowercase) == 1 assert len(packed_examples[0]) == sum(len(__lowercase) for x in orig_examples) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE , reason='''This test requires fairseq''') def UpperCamelCase__ ( self) -> List[Any]: if not FAIRSEQ_AVAILABLE: return __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = self._get_dataset(max_len=64) __UpperCamelCase :Union[str, Any] = 64 __UpperCamelCase :Tuple = ds.make_dynamic_sampler(__lowercase , required_batch_size_multiple=__lowercase) __UpperCamelCase :List[str] = [len(__lowercase) for x in batch_sampler] assert len(set(__lowercase)) > 1 # it's not dynamic batch size if every batch is the same length assert sum(__lowercase) == len(__lowercase) # no dropped or added examples __UpperCamelCase :int = DataLoader(__lowercase , batch_sampler=__lowercase , collate_fn=ds.collate_fn , num_workers=2) __UpperCamelCase :List[str] = [] __UpperCamelCase :int = [] for batch in data_loader: __UpperCamelCase :List[Any] = batch['''input_ids'''].shape __UpperCamelCase :Dict = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple __UpperCamelCase :Optional[int] = np.product(batch['''input_ids'''].shape) num_src_per_batch.append(__lowercase) if num_src_tokens > (max_tokens * 1.1): failures.append(__lowercase) assert num_src_per_batch[0] == max(__lowercase) if failures: raise AssertionError(f"""too many tokens in {len(__lowercase)} batches""") def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[Any] = self._get_dataset(max_len=512) __UpperCamelCase :Any = 2 __UpperCamelCase :List[Any] = ds.make_sortish_sampler(__lowercase , shuffle=__lowercase) __UpperCamelCase :List[Any] = DataLoader(__lowercase , batch_size=__lowercase , collate_fn=ds.collate_fn , num_workers=2) __UpperCamelCase :Tuple = DataLoader(__lowercase , batch_size=__lowercase , collate_fn=ds.collate_fn , num_workers=2 , sampler=__lowercase) __UpperCamelCase :int = tokenizer.pad_token_id def count_pad_tokens(__lowercase , __lowercase="input_ids"): return [batch[k].eq(__lowercase).sum().item() for batch in data_loader] assert sum(count_pad_tokens(__lowercase , k='''labels''')) < sum(count_pad_tokens(__lowercase , k='''labels''')) assert sum(count_pad_tokens(__lowercase)) < sum(count_pad_tokens(__lowercase)) assert len(__lowercase) == len(__lowercase) def UpperCamelCase__ ( self , __lowercase=1_000 , __lowercase=128) -> List[Any]: if os.getenv('''USE_REAL_DATA''' , __lowercase): __UpperCamelCase :Optional[Any] = '''examples/seq2seq/wmt_en_ro''' __UpperCamelCase :Dict = max_len * 2 * 64 if not Path(__lowercase).joinpath('''train.len''').exists(): save_len_file(__lowercase , __lowercase) else: __UpperCamelCase :Union[str, Any] = '''examples/seq2seq/test_data/wmt_en_ro''' __UpperCamelCase :Optional[int] = max_len * 4 save_len_file(__lowercase , __lowercase) __UpperCamelCase :str = AutoTokenizer.from_pretrained(__lowercase) __UpperCamelCase :List[Any] = SeqaSeqDataset( __lowercase , data_dir=__lowercase , type_path='''train''' , max_source_length=__lowercase , max_target_length=__lowercase , n_obs=__lowercase , ) return ds, max_tokens, tokenizer def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :List[Any] = self._get_dataset() __UpperCamelCase :List[str] = set(DistributedSortishSampler(__lowercase , 256 , num_replicas=2 , rank=0 , add_extra_examples=__lowercase)) __UpperCamelCase :Tuple = set(DistributedSortishSampler(__lowercase , 256 , num_replicas=2 , rank=1 , add_extra_examples=__lowercase)) assert idsa.intersection(__lowercase) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) def UpperCamelCase__ ( self , __lowercase) -> List[Any]: __UpperCamelCase :List[Any] = AutoTokenizer.from_pretrained(__lowercase , use_fast=__lowercase) if tok_name == MBART_TINY: __UpperCamelCase :Optional[Any] = SeqaSeqDataset( __lowercase , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir()) , type_path='''train''' , max_source_length=4 , max_target_length=8 , src_lang='''EN''' , tgt_lang='''FR''' , ) __UpperCamelCase :Tuple = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: __UpperCamelCase :Tuple = SeqaSeqDataset( __lowercase , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir()) , type_path='''train''' , max_source_length=4 , max_target_length=8 , ) __UpperCamelCase :Optional[int] = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(__lowercase) == 1 if tok_name == BART_TINY else len(__lowercase) == 0
105
0
from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = DistilBertTokenizer UpperCAmelCase__ = DistilBertTokenizerFast UpperCAmelCase__ = True @slow def A_ ( self : str ) -> Any: lowerCamelCase__ : List[str] = DistilBertTokenizer.from_pretrained('distilbert-base-uncased' ) lowerCamelCase__ : Dict = tokenizer.encode('sequence builders' , add_special_tokens=UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = tokenizer.encode('multi-sequence build' , add_special_tokens=UpperCAmelCase ) lowerCamelCase__ : List[str] = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase ) lowerCamelCase__ : List[Any] = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase , UpperCAmelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
50
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> list[tuple[int, int]]: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] = position lowerCamelCase__ : Optional[Any] = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] lowerCamelCase__ : Dict = [] for position in positions: lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_UpperCAmelCase ) return permissible_positions def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: return not any(elem == 0 for row in board for elem in row ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> bool: if is_complete(_UpperCAmelCase ): return True for position in get_valid_pos(_UpperCAmelCase , len(_UpperCAmelCase ) ): lowerCamelCase__ , lowerCamelCase__ : Optional[int] = position if board[y][x] == 0: lowerCamelCase__ : List[Any] = curr + 1 if open_knight_tour_helper(_UpperCAmelCase , _UpperCAmelCase , curr + 1 ): return True lowerCamelCase__ : Optional[Any] = 0 return False def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[list[int]]: lowerCamelCase__ : Any = [[0 for i in range(_UpperCAmelCase )] for j in range(_UpperCAmelCase )] for i in range(_UpperCAmelCase ): for j in range(_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = 1 if open_knight_tour_helper(_UpperCAmelCase , (i, j) , 1 ): return board lowerCamelCase__ : Optional[Any] = 0 lowerCamelCase__ : Any = F"""Open Kight Tour cannot be performed on a board of size {n}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
50
1
def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ ): try: __lowerCamelCase : Tuple = float(SCREAMING_SNAKE_CASE__ ) except ValueError: raise ValueError('Please enter a valid number' ) __lowerCamelCase : List[str] = decimal - int(SCREAMING_SNAKE_CASE__ ) if fractional_part == 0: return int(SCREAMING_SNAKE_CASE__ ), 1 else: __lowerCamelCase : List[Any] = len(str(SCREAMING_SNAKE_CASE__ ).split('.' )[1] ) __lowerCamelCase : Optional[int] = int(decimal * (10**number_of_frac_digits) ) __lowerCamelCase : List[str] = 10**number_of_frac_digits __lowerCamelCase , __lowerCamelCase : Union[str, Any] = denominator, numerator while True: __lowerCamelCase : str = dividend % divisor if remainder == 0: break __lowerCamelCase , __lowerCamelCase : Any = divisor, remainder __lowerCamelCase , __lowerCamelCase : Dict = numerator / divisor, denominator / divisor return int(SCREAMING_SNAKE_CASE__ ), int(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": print(F"""{decimal_to_fraction(2) = }""") print(F"""{decimal_to_fraction(89.0) = }""") print(F"""{decimal_to_fraction('67') = }""") print(F"""{decimal_to_fraction('45.0') = }""") print(F"""{decimal_to_fraction(1.5) = }""") print(F"""{decimal_to_fraction('6.25') = }""") print(F"""{decimal_to_fraction('78td') = }""")
194
from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def UpperCamelCase__ ( SCREAMING_SNAKE_CASE__ ): __lowerCamelCase : Dict = prime_factors(SCREAMING_SNAKE_CASE__ ) if is_square_free(SCREAMING_SNAKE_CASE__ ): return -1 if len(SCREAMING_SNAKE_CASE__ ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
194
1
'''simple docstring''' import socket def UpperCamelCase ( ): A__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) A__ = socket.gethostname() A__ = 1_23_12 sock.connect((host, port) ) sock.send(b"Hello server!" ) with open("Received_file" , "wb" ) as out_file: print("File opened" ) print("Receiving data..." ) while True: A__ = sock.recv(10_24 ) if not data: break out_file.write(_lowerCamelCase ) print("Successfully received the file" ) sock.close() print("Connection closed" ) if __name__ == "__main__": main()
237
'''simple docstring''' import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) __lowerCAmelCase : List[str] ={ "sample_size": 32, "in_channels": 3, "out_channels": 3, "layers_per_block": 2, "num_class_embeds": 1000, "block_out_channels": [32, 64], "attention_head_dim": 8, "down_block_types": [ "ResnetDownsampleBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "scale_shift", "upsample_type": "resnet", "downsample_type": "resnet", } __lowerCAmelCase : Dict ={ "sample_size": 64, "in_channels": 3, "out_channels": 3, "layers_per_block": 3, "num_class_embeds": 1000, "block_out_channels": [192, 192 * 2, 192 * 3, 192 * 4], "attention_head_dim": 64, "down_block_types": [ "ResnetDownsampleBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "AttnUpBlock2D", "AttnUpBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "scale_shift", "upsample_type": "resnet", "downsample_type": "resnet", } __lowerCAmelCase : Union[str, Any] ={ "sample_size": 256, "in_channels": 3, "out_channels": 3, "layers_per_block": 2, "num_class_embeds": None, "block_out_channels": [256, 256, 256 * 2, 256 * 2, 256 * 4, 256 * 4], "attention_head_dim": 64, "down_block_types": [ "ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "AttnUpBlock2D", "AttnUpBlock2D", "ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "default", "upsample_type": "resnet", "downsample_type": "resnet", } __lowerCAmelCase : str ={ "num_train_timesteps": 40, "sigma_min": 0.002, "sigma_max": 80.0, } __lowerCAmelCase : Tuple ={ "num_train_timesteps": 201, "sigma_min": 0.002, "sigma_max": 80.0, } __lowerCAmelCase : Dict ={ "num_train_timesteps": 151, "sigma_min": 0.002, "sigma_max": 80.0, } def UpperCamelCase ( _lowerCamelCase : Tuple ): if isinstance(_lowerCamelCase , _lowerCamelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("boolean value expected" ) def UpperCamelCase ( _lowerCamelCase : List[str] , _lowerCamelCase : List[str] , _lowerCamelCase : Optional[int] , _lowerCamelCase : Dict , _lowerCamelCase : Optional[int]=False ): A__ = checkpoint[F"{old_prefix}.in_layers.0.weight"] A__ = checkpoint[F"{old_prefix}.in_layers.0.bias"] A__ = checkpoint[F"{old_prefix}.in_layers.2.weight"] A__ = checkpoint[F"{old_prefix}.in_layers.2.bias"] A__ = checkpoint[F"{old_prefix}.emb_layers.1.weight"] A__ = checkpoint[F"{old_prefix}.emb_layers.1.bias"] A__ = checkpoint[F"{old_prefix}.out_layers.0.weight"] A__ = checkpoint[F"{old_prefix}.out_layers.0.bias"] A__ = checkpoint[F"{old_prefix}.out_layers.3.weight"] A__ = checkpoint[F"{old_prefix}.out_layers.3.bias"] if has_skip: A__ = checkpoint[F"{old_prefix}.skip_connection.weight"] A__ = checkpoint[F"{old_prefix}.skip_connection.bias"] return new_checkpoint def UpperCamelCase ( _lowerCamelCase : str , _lowerCamelCase : Any , _lowerCamelCase : Tuple , _lowerCamelCase : Optional[int] , _lowerCamelCase : List[Any]=None ): A__, A__, A__ = checkpoint[F"{old_prefix}.qkv.weight"].chunk(3 , dim=0 ) A__, A__, A__ = checkpoint[F"{old_prefix}.qkv.bias"].chunk(3 , dim=0 ) A__ = checkpoint[F"{old_prefix}.norm.weight"] A__ = checkpoint[F"{old_prefix}.norm.bias"] A__ = weight_q.squeeze(-1 ).squeeze(-1 ) A__ = bias_q.squeeze(-1 ).squeeze(-1 ) A__ = weight_k.squeeze(-1 ).squeeze(-1 ) A__ = bias_k.squeeze(-1 ).squeeze(-1 ) A__ = weight_v.squeeze(-1 ).squeeze(-1 ) A__ = bias_v.squeeze(-1 ).squeeze(-1 ) A__ = ( checkpoint[F"{old_prefix}.proj_out.weight"].squeeze(-1 ).squeeze(-1 ) ) A__ = checkpoint[F"{old_prefix}.proj_out.bias"].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def UpperCamelCase ( _lowerCamelCase : str , _lowerCamelCase : List[str] ): A__ = torch.load(_lowerCamelCase , map_location="cpu" ) A__ = {} A__ = checkpoint["time_embed.0.weight"] A__ = checkpoint["time_embed.0.bias"] A__ = checkpoint["time_embed.2.weight"] A__ = checkpoint["time_embed.2.bias"] if unet_config["num_class_embeds"] is not None: A__ = checkpoint["label_emb.weight"] A__ = checkpoint["input_blocks.0.0.weight"] A__ = checkpoint["input_blocks.0.0.bias"] A__ = unet_config["down_block_types"] A__ = unet_config["layers_per_block"] A__ = unet_config["attention_head_dim"] A__ = unet_config["block_out_channels"] A__ = 1 A__ = channels_list[0] for i, layer_type in enumerate(_lowerCamelCase ): A__ = channels_list[i] A__ = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(_lowerCamelCase ): A__ = F"down_blocks.{i}.resnets.{j}" A__ = F"input_blocks.{current_layer}.0" A__ = True if j == 0 and downsample_block_has_skip else False A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , has_skip=_lowerCamelCase ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(_lowerCamelCase ): A__ = F"down_blocks.{i}.resnets.{j}" A__ = F"input_blocks.{current_layer}.0" A__ = True if j == 0 and downsample_block_has_skip else False A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , has_skip=_lowerCamelCase ) A__ = F"down_blocks.{i}.attentions.{j}" A__ = F"input_blocks.{current_layer}.1" A__ = convert_attention( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) current_layer += 1 if i != len(_lowerCamelCase ) - 1: A__ = F"down_blocks.{i}.downsamplers.0" A__ = F"input_blocks.{current_layer}.0" A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) current_layer += 1 A__ = current_channels # hardcoded the mid-block for now A__ = "mid_block.resnets.0" A__ = "middle_block.0" A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A__ = "mid_block.attentions.0" A__ = "middle_block.1" A__ = convert_attention(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A__ = "mid_block.resnets.1" A__ = "middle_block.2" A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A__ = 0 A__ = unet_config["up_block_types"] for i, layer_type in enumerate(_lowerCamelCase ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): A__ = F"up_blocks.{i}.resnets.{j}" A__ = F"output_blocks.{current_layer}.0" A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , has_skip=_lowerCamelCase ) current_layer += 1 if i != len(_lowerCamelCase ) - 1: A__ = F"up_blocks.{i}.upsamplers.0" A__ = F"output_blocks.{current_layer-1}.1" A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): A__ = F"up_blocks.{i}.resnets.{j}" A__ = F"output_blocks.{current_layer}.0" A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , has_skip=_lowerCamelCase ) A__ = F"up_blocks.{i}.attentions.{j}" A__ = F"output_blocks.{current_layer}.1" A__ = convert_attention( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) current_layer += 1 if i != len(_lowerCamelCase ) - 1: A__ = F"up_blocks.{i}.upsamplers.0" A__ = F"output_blocks.{current_layer-1}.2" A__ = convert_resnet(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A__ = checkpoint["out.0.weight"] A__ = checkpoint["out.0.bias"] A__ = checkpoint["out.2.weight"] A__ = checkpoint["out.2.bias"] return new_checkpoint if __name__ == "__main__": __lowerCAmelCase : List[Any] =argparse.ArgumentParser() parser.add_argument("--unet_path", default=None, type=str, required=True, help="Path to the unet.pt to convert.") parser.add_argument( "--dump_path", default=None, type=str, required=True, help="Path to output the converted UNet model." ) parser.add_argument("--class_cond", default=True, type=str, help="Whether the model is class-conditional.") __lowerCAmelCase : Optional[Any] =parser.parse_args() __lowerCAmelCase : List[Any] =strabool(args.class_cond) __lowerCAmelCase : List[str] =os.path.basename(args.unet_path) print(f"""Checkpoint: {ckpt_name}""") # Get U-Net config if "imagenet64" in ckpt_name: __lowerCAmelCase : List[str] =IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): __lowerCAmelCase : List[str] =LSUN_256_UNET_CONFIG elif "test" in ckpt_name: __lowerCAmelCase : Any =TEST_UNET_CONFIG else: raise ValueError(f"""Checkpoint type {ckpt_name} is not currently supported.""") if not args.class_cond: __lowerCAmelCase : Dict =None __lowerCAmelCase : Optional[int] =con_pt_to_diffuser(args.unet_path, unet_config) __lowerCAmelCase : Dict =UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: __lowerCAmelCase : List[str] =CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: __lowerCAmelCase : Dict =CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): __lowerCAmelCase : Dict =CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(f"""Checkpoint type {ckpt_name} is not currently supported.""") __lowerCAmelCase : Dict =CMStochasticIterativeScheduler(**scheduler_config) __lowerCAmelCase : str =ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
237
1
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def snake_case_(_UpperCamelCase ): """simple docstring""" _snake_case = checkpoints.load_tax_checkpoint(_UpperCamelCase ) _snake_case = flatten_dict(_UpperCamelCase ) return flax_params def snake_case_(_UpperCamelCase ): """simple docstring""" _snake_case = {} _snake_case = { '''token_embedder''': '''embeddings''', '''encoder_norm''': '''layernorm''', '''kernel''': '''weight''', '''.out''': '''.output''', '''scale''': '''weight''', '''embedders_0.pos_embedding''': '''row_embedder.weight''', '''embedders_1.pos_embedding''': '''column_embedder.weight''', } _snake_case = { '''query''': '''attention.query''', '''key''': '''attention.key''', '''value''': '''attention.value''', '''output.dense''': '''output''', '''encoder_decoder_attention.o''': '''encoder_decoder_attention.attention.o''', '''pre_self_attention_layer_norm''': '''self_attention.layer_norm''', '''pre_cross_attention_layer_norm''': '''encoder_decoder_attention.layer_norm''', '''mlp.''': '''mlp.DenseReluDense.''', '''pre_mlp_layer_norm''': '''mlp.layer_norm''', '''self_attention.o''': '''self_attention.attention.o''', '''decoder.embeddings.embedding''': '''decoder.embed_tokens.weight''', '''decoder.relpos_bias.rel_embedding''': '''decoder.layer.0.self_attention.attention.relative_attention_bias.weight''', '''decoder.decoder_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.logits_dense.weight''': '''decoder.lm_head.weight''', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key _snake_case = '''.'''.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): _snake_case = new_key.replace(_UpperCamelCase , _UpperCamelCase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): _snake_case = new_key.replace(_UpperCamelCase , _UpperCamelCase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number _snake_case = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , _UpperCamelCase ) _snake_case = new_key.replace('''encoder''' , '''encoder.encoder''' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number _snake_case = re.sub(R'''layers_(\d+)''' , R'''layer.\1''' , _UpperCamelCase ) _snake_case = flax_dict[key] _snake_case = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): _snake_case = torch.from_numpy(converted_dict[key].T ) else: _snake_case = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def snake_case_(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase=False , _UpperCamelCase=False ): """simple docstring""" _snake_case = get_flax_param(_UpperCamelCase ) if not use_large: _snake_case = PixaStructVisionConfig() _snake_case = PixaStructTextConfig() else: _snake_case = PixaStructVisionConfig( hidden_size=1_536 , d_ff=3_968 , num_attention_heads=24 , num_hidden_layers=18 ) _snake_case = PixaStructTextConfig(hidden_size=1_536 , d_ff=3_968 , num_heads=24 , num_layers=18 ) _snake_case = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=_UpperCamelCase ) _snake_case = PixaStructForConditionalGeneration(_UpperCamelCase ) _snake_case = rename_and_convert_flax_params(_UpperCamelCase ) model.load_state_dict(_UpperCamelCase ) _snake_case = AutoTokenizer.from_pretrained('''ybelkada/test-pix2struct-tokenizer''' ) _snake_case = PixaStructImageProcessor() _snake_case = PixaStructProcessor(image_processor=_UpperCamelCase , tokenizer=_UpperCamelCase ) if use_large: _snake_case = 4_096 _snake_case = True # mkdir if needed os.makedirs(_UpperCamelCase , exist_ok=_UpperCamelCase ) model.save_pretrained(_UpperCamelCase ) processor.save_pretrained(_UpperCamelCase ) print('''Model saved in {}'''.format(_UpperCamelCase ) ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') __A = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
356
import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ : def __init__( self : Optional[Any] , A__ : str , A__ : Any=13 , A__ : str=[30, 30] , A__ : int=2 , A__ : Dict=3 , A__ : str=True , A__ : Union[str, Any]=True , A__ : Any=32 , A__ : int=5 , A__ : str=4 , A__ : List[Any]=37 , A__ : Union[str, Any]="gelu" , A__ : Dict=0.1 , A__ : Dict=0.1 , A__ : Tuple=10 , A__ : Dict=0.02 , A__ : Any=3 , A__ : Union[str, Any]=None , A__ : Optional[Any]=8 , A__ : Dict=10 , ) -> Optional[Any]: _snake_case = parent _snake_case = batch_size _snake_case = image_size _snake_case = patch_size _snake_case = num_channels _snake_case = is_training _snake_case = use_labels _snake_case = hidden_size _snake_case = num_hidden_layers _snake_case = num_attention_heads _snake_case = intermediate_size _snake_case = hidden_act _snake_case = hidden_dropout_prob _snake_case = attention_probs_dropout_prob _snake_case = type_sequence_label_size _snake_case = initializer_range _snake_case = num_labels _snake_case = scope _snake_case = n_targets _snake_case = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens _snake_case = (image_size[1] // patch_size) * (image_size[0] // patch_size) _snake_case = num_patches + 1 + self.num_detection_tokens def UpperCamelCase_ ( self : List[str] ) -> str: _snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) _snake_case = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) _snake_case = [] for i in range(self.batch_size ): _snake_case = {} _snake_case = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=A__ ) _snake_case = torch.rand(self.n_targets , 4 , device=A__ ) labels.append(A__ ) _snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self : Dict ) -> List[Any]: return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A__ , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def UpperCamelCase_ ( self : Any , A__ : Any , A__ : str , A__ : Tuple ) -> Dict: _snake_case = YolosModel(config=A__ ) model.to(A__ ) model.eval() _snake_case = model(A__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def UpperCamelCase_ ( self : Dict , A__ : List[str] , A__ : Optional[Any] , A__ : str ) -> int: _snake_case = YolosForObjectDetection(A__ ) model.to(A__ ) model.eval() _snake_case = model(pixel_values=A__ ) _snake_case = model(A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) _snake_case = model(pixel_values=A__ , labels=A__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def UpperCamelCase_ ( self : Optional[Any] ) -> Tuple: _snake_case = self.prepare_config_and_inputs() _snake_case, _snake_case, _snake_case = config_and_inputs _snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ ( __lowercase , __lowercase , unittest.TestCase ): UpperCamelCase_ : Optional[int] = (YolosModel, YolosForObjectDetection) if is_torch_available() else () UpperCamelCase_ : int = ( {"feature-extraction": YolosModel, "object-detection": YolosForObjectDetection} if is_torch_available() else {} ) UpperCamelCase_ : List[str] = False UpperCamelCase_ : List[Any] = False UpperCamelCase_ : List[Any] = False UpperCamelCase_ : Tuple = False def UpperCamelCase_ ( self : Dict , A__ : List[Any] , A__ : List[str] , A__ : Optional[int]=False ) -> Optional[int]: _snake_case = super()._prepare_for_class(A__ , A__ , return_labels=A__ ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": _snake_case = [] for i in range(self.model_tester.batch_size ): _snake_case = {} _snake_case = torch.ones( size=(self.model_tester.n_targets,) , device=A__ , dtype=torch.long ) _snake_case = torch.ones( self.model_tester.n_targets , 4 , device=A__ , dtype=torch.float ) labels.append(A__ ) _snake_case = labels return inputs_dict def UpperCamelCase_ ( self : List[Any] ) -> List[str]: _snake_case = YolosModelTester(self ) _snake_case = ConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=37 ) def UpperCamelCase_ ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() def UpperCamelCase_ ( self : List[Any] ) -> str: # YOLOS does not use inputs_embeds pass def UpperCamelCase_ ( self : Union[str, Any] ) -> List[str]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(A__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A__ , nn.Linear ) ) def UpperCamelCase_ ( self : List[Any] ) -> Optional[Any]: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = model_class(A__ ) _snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case = [*signature.parameters.keys()] _snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A__ ) def UpperCamelCase_ ( self : List[str] ) -> List[Any]: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def UpperCamelCase_ ( self : Union[str, Any] ) -> int: _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() _snake_case = True # in YOLOS, the seq_len is different _snake_case = self.model_tester.expected_seq_len for model_class in self.all_model_classes: _snake_case = True _snake_case = False _snake_case = True _snake_case = model_class(A__ ) model.to(A__ ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(A__ , A__ ) ) _snake_case = outputs.attentions self.assertEqual(len(A__ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _snake_case = True _snake_case = model_class(A__ ) model.to(A__ ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(A__ , A__ ) ) _snake_case = outputs.attentions self.assertEqual(len(A__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) _snake_case = len(A__ ) # Check attention is always last and order is fine _snake_case = True _snake_case = True _snake_case = model_class(A__ ) model.to(A__ ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(A__ , A__ ) ) _snake_case = 1 self.assertEqual(out_len + added_hidden_states , len(A__ ) ) _snake_case = outputs.attentions self.assertEqual(len(A__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def UpperCamelCase_ ( self : int ) -> Dict: def check_hidden_states_output(A__ : Optional[int] , A__ : Union[str, Any] , A__ : int ): _snake_case = model_class(A__ ) model.to(A__ ) model.eval() with torch.no_grad(): _snake_case = model(**self._prepare_for_class(A__ , A__ ) ) _snake_case = outputs.hidden_states _snake_case = getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A__ ) , A__ ) # YOLOS has a different seq_length _snake_case = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) _snake_case, _snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case = True check_hidden_states_output(A__ , A__ , A__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case = True check_hidden_states_output(A__ , A__ , A__ ) def UpperCamelCase_ ( self : Optional[Any] ) -> str: _snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*A__ ) @slow def UpperCamelCase_ ( self : List[str] ) -> Dict: for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case = YolosModel.from_pretrained(A__ ) self.assertIsNotNone(A__ ) def snake_case_() -> str: """simple docstring""" _snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : Any ) -> str: return AutoImageProcessor.from_pretrained('''hustvl/yolos-small''' ) if is_vision_available() else None @slow def UpperCamelCase_ ( self : Tuple ) -> str: _snake_case = YolosForObjectDetection.from_pretrained('''hustvl/yolos-small''' ).to(A__ ) _snake_case = self.default_image_processor _snake_case = prepare_img() _snake_case = image_processor(images=A__ , return_tensors='''pt''' ).to(A__ ) # forward pass with torch.no_grad(): _snake_case = model(inputs.pixel_values ) # verify outputs _snake_case = torch.Size((1, 100, 92) ) self.assertEqual(outputs.logits.shape , A__ ) _snake_case = torch.tensor( [[-24.0248, -10.3024, -14.8290], [-42.0392, -16.8200, -27.4334], [-27.2743, -11.8154, -18.7148]] , device=A__ , ) _snake_case = torch.tensor( [[0.2559, 0.5455, 0.4706], [0.2989, 0.7279, 0.1875], [0.7732, 0.4017, 0.4462]] , device=A__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , A__ , atol=1e-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , A__ , atol=1e-4 ) ) # verify postprocessing _snake_case = image_processor.post_process_object_detection( A__ , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] _snake_case = torch.tensor([0.9994, 0.9790, 0.9964, 0.9972, 0.9861] ).to(A__ ) _snake_case = [75, 75, 17, 63, 17] _snake_case = torch.tensor([335.0609, 79.3848, 375.4216, 187.2495] ).to(A__ ) self.assertEqual(len(results['''scores'''] ) , 5 ) self.assertTrue(torch.allclose(results['''scores'''] , A__ , atol=1e-4 ) ) self.assertSequenceEqual(results['''labels'''].tolist() , A__ ) self.assertTrue(torch.allclose(results['''boxes'''][0, :] , A__ ) )
278
0
'''simple docstring''' import math def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : int ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5, int(math.sqrt(_snake_case ) + 1 ), 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ : float = 0.1 ) -> int: UpperCAmelCase_ : Tuple = 3 UpperCAmelCase_ : Any = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1, (j + 2) * (j + 2), j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
125
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def lowercase ( _snake_case : Any ) ->Union[str, Any]: """simple docstring""" __snake_case : Tuple = [2, 2, 6, 2] if '''tiny''' in model_name else [2, 2, 18, 2] __snake_case : Dict = True if '''large''' in model_name or '''huge''' in model_name else False __snake_case : Optional[int] = True if '''large''' in model_name or '''huge''' in model_name else False __snake_case : Optional[int] = True if '''large''' in model_name or '''huge''' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: __snake_case : Tuple = [3, 3, 3, 3] __snake_case : Dict = [5, 5, 5, 5] elif "fl4" in model_name: __snake_case : Any = [4, 4, 4, 4] __snake_case : List[str] = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: __snake_case : Optional[int] = [3, 3, 3, 3] if "lrf" in model_name: __snake_case : Any = [3, 3, 3, 3] else: __snake_case : int = [2, 2, 2, 2] if "tiny" in model_name: __snake_case : str = 96 elif "small" in model_name: __snake_case : Optional[int] = 96 elif "base" in model_name: __snake_case : Any = 128 elif "large" in model_name: __snake_case : Optional[Any] = 192 elif "xlarge" in model_name: __snake_case : List[Any] = 256 elif "huge" in model_name: __snake_case : Union[str, Any] = 352 # set label information __snake_case : Union[str, Any] = '''huggingface/label-files''' if "large" in model_name or "huge" in model_name: __snake_case : int = '''imagenet-22k-id2label.json''' else: __snake_case : Optional[Any] = '''imagenet-1k-id2label.json''' __snake_case : int = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type='''dataset''' ) , '''r''' ) ) __snake_case : Dict = {int(_snake_case ): v for k, v in idalabel.items()} __snake_case : Optional[int] = {v: k for k, v in idalabel.items()} __snake_case : Optional[Any] = FocalNetConfig( embed_dim=_snake_case , depths=_snake_case , focal_levels=_snake_case , focal_windows=_snake_case , use_conv_embed=_snake_case , idalabel=_snake_case , labelaid=_snake_case , use_post_layernorm=_snake_case , use_layerscale=_snake_case , ) return config def lowercase ( _snake_case : Dict ) ->List[Any]: """simple docstring""" if "patch_embed.proj" in name: __snake_case : Tuple = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: __snake_case : Tuple = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if "layers" in name: __snake_case : List[Any] = '''encoder.''' + name if "encoder.layers" in name: __snake_case : Optional[Any] = name.replace('''encoder.layers''' , '''encoder.stages''' ) if "downsample.proj" in name: __snake_case : Any = name.replace('''downsample.proj''' , '''downsample.projection''' ) if "blocks" in name: __snake_case : List[str] = name.replace('''blocks''' , '''layers''' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: __snake_case : Any = name.replace('''modulation.f''' , '''modulation.projection_in''' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: __snake_case : List[Any] = name.replace('''modulation.h''' , '''modulation.projection_context''' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: __snake_case : int = name.replace('''modulation.proj''' , '''modulation.projection_out''' ) if name == "norm.weight": __snake_case : Optional[Any] = '''layernorm.weight''' if name == "norm.bias": __snake_case : List[str] = '''layernorm.bias''' if "head" in name: __snake_case : Union[str, Any] = name.replace('''head''' , '''classifier''' ) else: __snake_case : int = '''focalnet.''' + name return name def lowercase ( _snake_case : Tuple , _snake_case : Dict , _snake_case : List[str]=False ) ->Any: """simple docstring""" __snake_case : List[Any] = { '''focalnet-tiny''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth''', '''focalnet-tiny-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth''', '''focalnet-small''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth''', '''focalnet-small-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth''', '''focalnet-base''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth''', '''focalnet-base-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth''', '''focalnet-large-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth''', '''focalnet-large-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth''', '''focalnet-xlarge-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth''', '''focalnet-xlarge-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth''', } # fmt: on __snake_case : int = model_name_to_url[model_name] print('''Checkpoint URL: ''' , _snake_case ) __snake_case : int = torch.hub.load_state_dict_from_url(_snake_case , map_location='''cpu''' )['''model'''] # rename keys for key in state_dict.copy().keys(): __snake_case : str = state_dict.pop(_snake_case ) __snake_case : Tuple = val __snake_case : Any = get_focalnet_config(_snake_case ) __snake_case : List[Any] = FocalNetForImageClassification(_snake_case ) model.eval() # load state dict model.load_state_dict(_snake_case ) # verify conversion __snake_case : str = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __snake_case : Any = BitImageProcessor( do_resize=_snake_case , size={'''shortest_edge''': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=_snake_case , crop_size=224 , do_normalize=_snake_case , image_mean=_snake_case , image_std=_snake_case , ) __snake_case : List[str] = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) __snake_case : int = processor(images=_snake_case , return_tensors='''pt''' ) __snake_case : Optional[int] = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) __snake_case : Optional[Any] = image_transforms(_snake_case ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , _snake_case , atol=1e-4 ) __snake_case : Tuple = model(**_snake_case ) __snake_case : str = outputs.logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) print('''First values of logits:''' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": __snake_case : Any = torch.tensor([0.2166, -0.4368, 0.2191] ) elif model_name == "focalnet-tiny-lrf": __snake_case : int = torch.tensor([1.1669, 0.0125, -0.1695] ) elif model_name == "focalnet-small": __snake_case : Optional[int] = torch.tensor([0.4917, -0.0430, 0.1341] ) elif model_name == "focalnet-small-lrf": __snake_case : List[Any] = torch.tensor([-0.2588, -0.5342, -0.2331] ) elif model_name == "focalnet-base": __snake_case : Union[str, Any] = torch.tensor([-0.1655, -0.4090, -0.1730] ) elif model_name == "focalnet-base-lrf": __snake_case : List[str] = torch.tensor([0.5306, -0.0483, -0.3928] ) assert torch.allclose(outputs.logits[0, :3] , _snake_case , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f"""Saving model and processor of {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_snake_case ) processor.save_pretrained(_snake_case ) if push_to_hub: print(f"""Pushing model and processor of {model_name} to the hub...""" ) model.push_to_hub(f"""{model_name}""" ) processor.push_to_hub(f"""{model_name}""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""focalnet-tiny""", type=str, help="""Name of the FocalNet model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub.""", ) SCREAMING_SNAKE_CASE : str = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
102
0
import argparse import importlib from pathlib import Path # Test all the extensions added in the setup snake_case__ : int = [ 'kernels/rwkv/wkv_cuda.cu', 'kernels/rwkv/wkv_op.cpp', 'kernels/deformable_detr/ms_deform_attn.h', 'kernels/deformable_detr/cuda/ms_deform_im2col_cuda.cuh', 'models/graphormer/algos_graphormer.pyx', ] def _a ( lowerCamelCase: Dict ) -> List[Any]: '''simple docstring''' for file in FILES_TO_FIND: if not (transformers_path / file).exists(): return False return True if __name__ == "__main__": snake_case__ : Tuple = argparse.ArgumentParser() parser.add_argument('--check_lib', action='store_true', help='Whether to check the build or the actual package.') snake_case__ : List[str] = parser.parse_args() if args.check_lib: snake_case__ : List[Any] = importlib.import_module('transformers') snake_case__ : Tuple = Path(transformers_module.__file__).parent else: snake_case__ : List[str] = Path.cwd() / 'build/lib/transformers' if not test_custom_files_are_present(transformers_path): raise ValueError('The built release does not contain the custom files. Fix this before going further!')
359
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def _a ( lowerCamelCase: List[Any] ) -> List[Any]: # picklable for multiprocessing '''simple docstring''' return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def _a ( ) -> Dict: '''simple docstring''' with parallel_backend('''spark''' ): assert ParallelBackendConfig.backend_name == "spark" __A = [1, 2, 3] with pytest.raises(lowerCamelCase ): with parallel_backend('''unsupported backend''' ): map_nested(lowerCamelCase , lowerCamelCase , num_proc=2 ) with pytest.raises(lowerCamelCase ): with parallel_backend('''unsupported backend''' ): map_nested(lowerCamelCase , lowerCamelCase , num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize('''num_proc''' , [2, -1] ) def _a ( lowerCamelCase: Optional[int] ) -> List[str]: '''simple docstring''' __A = [1, 2] __A = {'''a''': 1, '''b''': 2} __A = {'''a''': [1, 2], '''b''': [3, 4]} __A = {'''a''': {'''1''': 1}, '''b''': 2} __A = {'''a''': 1, '''b''': 2, '''c''': 3, '''d''': 4} __A = [2, 3] __A = {'''a''': 2, '''b''': 3} __A = {'''a''': [2, 3], '''b''': [4, 5]} __A = {'''a''': {'''1''': 2}, '''b''': 3} __A = {'''a''': 2, '''b''': 3, '''c''': 4, '''d''': 5} with parallel_backend('''spark''' ): assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa assert map_nested(lowerCamelCase , lowerCamelCase , num_proc=lowerCamelCase ) == expected_map_nested_sa
250
0
def SCREAMING_SNAKE_CASE__ ( __a ): if not head: return True # split the list to two parts snake_case_ ,snake_case_ : Optional[Any] = head.next, head while fast and fast.next: snake_case_ : Optional[int] = fast.next.next snake_case_ : int = slow.next snake_case_ : int = slow.next snake_case_ : Tuple = None # Don't forget here! But forget still works! # reverse the second part snake_case_ : Dict = None while second: snake_case_ : Any = second.next snake_case_ : Tuple = node snake_case_ : str = second snake_case_ : Dict = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False snake_case_ : Any = node.next snake_case_ : Dict = head.next return True def SCREAMING_SNAKE_CASE__ ( __a ): if not head or not head.next: return True # 1. Get the midpoint (slow) snake_case_ : Optional[Any] = head while fast and fast.next: snake_case_ ,snake_case_ : Tuple = fast.next.next, slow.next # 2. Push the second half into the stack snake_case_ : Union[str, Any] = [slow.val] while slow.next: snake_case_ : str = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False snake_case_ : int = cur.next return True def SCREAMING_SNAKE_CASE__ ( __a ): if not head or not head.next: return True snake_case_ : int = {} snake_case_ : Tuple = 0 while head: if head.val in d: d[head.val].append(__a ) else: snake_case_ : Tuple = [pos] snake_case_ : Union[str, Any] = head.next pos += 1 snake_case_ : Dict = pos - 1 snake_case_ : Dict = 0 for v in d.values(): if len(__a ) % 2 != 0: middle += 1 else: snake_case_ : Dict = 0 for i in range(0 , len(__a ) ): if v[i] + v[len(__a ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
327
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): def __init__( self : List[str] , _A : List[Any] , _A : bool = True , _A : Dict[str, int] = None , _A : int = 32 , _A : bool = True , _A : Union[int, float] = 1 / 255 , _A : bool = True , _A : bool = True , _A : Optional[Union[float, List[float]]] = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] , _A : Optional[Union[float, List[float]]] = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] , _A : bool = True , _A : Tuple=7 , _A : Tuple=30 , _A : int=400 , _A : Tuple=3 , ) -> Optional[int]: """simple docstring""" snake_case_ : str = parent snake_case_ : str = do_resize snake_case_ : str = size if size is not None else {'shortest_edge': 288} snake_case_ : Any = size_divisor snake_case_ : Any = do_rescale snake_case_ : Union[str, Any] = rescale_factor snake_case_ : str = do_normalize snake_case_ : int = do_center_crop snake_case_ : str = image_mean snake_case_ : int = image_std snake_case_ : Any = do_pad snake_case_ : Optional[int] = batch_size snake_case_ : List[str] = num_channels snake_case_ : Any = min_resolution snake_case_ : str = max_resolution def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def UpperCAmelCase_ ( self : Dict , _A : str , _A : Union[str, Any]=False ) -> int: """simple docstring""" if not batched: snake_case_ : Optional[int] = self.size['shortest_edge'] snake_case_ : List[Any] = image_inputs[0] if isinstance(_A , Image.Image ): snake_case_ ,snake_case_ : Optional[Any] = image.size else: snake_case_ ,snake_case_ : str = image.shape[1], image.shape[2] snake_case_ : Dict = size / min(_A , _A ) if h < w: snake_case_ ,snake_case_ : str = size, scale * w else: snake_case_ ,snake_case_ : Tuple = scale * h, size snake_case_ : Dict = int((1333 / 800) * size ) if max(_A , _A ) > max_size: snake_case_ : Union[str, Any] = max_size / max(_A , _A ) snake_case_ : Any = newh * scale snake_case_ : Union[str, Any] = neww * scale snake_case_ ,snake_case_ : Any = int(newh + 0.5 ), int(neww + 0.5 ) snake_case_ ,snake_case_ : int = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: snake_case_ : Optional[int] = [] for image in image_inputs: snake_case_ ,snake_case_ : Optional[int] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) snake_case_ : str = max(_A , key=lambda _A : item[0] )[0] snake_case_ : List[str] = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class SCREAMING_SNAKE_CASE_ ( snake_case_ , unittest.TestCase ): __magic_name__: List[Any] = BridgeTowerImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self : List[Any] ) -> Tuple: """simple docstring""" snake_case_ : int = BridgeTowerImageProcessingTester(self ) @property def UpperCAmelCase_ ( self : int ) -> Any: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: """simple docstring""" snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , 'image_mean' ) ) self.assertTrue(hasattr(_A , 'image_std' ) ) self.assertTrue(hasattr(_A , 'do_normalize' ) ) self.assertTrue(hasattr(_A , 'do_resize' ) ) self.assertTrue(hasattr(_A , 'size' ) ) self.assertTrue(hasattr(_A , 'size_divisor' ) ) def UpperCAmelCase_ ( self : Optional[int] ) -> List[str]: """simple docstring""" pass def UpperCAmelCase_ ( self : Tuple ) -> Tuple: """simple docstring""" snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input snake_case_ : Dict = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values snake_case_ ,snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : List[str] = image_processing(_A , return_tensors='pt' ).pixel_values snake_case_ ,snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: """simple docstring""" snake_case_ : int = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input snake_case_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values snake_case_ ,snake_case_ : Tuple = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : Any = image_processing(_A , return_tensors='pt' ).pixel_values snake_case_ ,snake_case_ : Any = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def UpperCAmelCase_ ( self : Dict ) -> List[str]: """simple docstring""" snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input snake_case_ : Optional[int] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values snake_case_ ,snake_case_ : Optional[Any] = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched snake_case_ : str = image_processing(_A , return_tensors='pt' ).pixel_values snake_case_ ,snake_case_ : Tuple = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
327
1
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { "salesforce/blip2-opt-2.7b": "https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json", } class _snake_case ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' A__ : int = "blip_2_vision_model" def __init__( self: List[str] ,lowerCamelCase_: int=1408 ,lowerCamelCase_: str=6144 ,lowerCamelCase_: str=39 ,lowerCamelCase_: List[str]=16 ,lowerCamelCase_: List[str]=224 ,lowerCamelCase_: List[Any]=14 ,lowerCamelCase_: Dict="gelu" ,lowerCamelCase_: int=0.0_0_0_0_1 ,lowerCamelCase_: Optional[int]=0.0 ,lowerCamelCase_: Optional[int]=1e-10 ,lowerCamelCase_: Dict=True ,**lowerCamelCase_: Optional[Any] ,) -> int: super().__init__(**lowerCamelCase_ ) UpperCAmelCase_ : Optional[int] = hidden_size UpperCAmelCase_ : str = intermediate_size UpperCAmelCase_ : Dict = num_hidden_layers UpperCAmelCase_ : Optional[Any] = num_attention_heads UpperCAmelCase_ : int = patch_size UpperCAmelCase_ : List[Any] = image_size UpperCAmelCase_ : Dict = initializer_range UpperCAmelCase_ : str = attention_dropout UpperCAmelCase_ : Dict = layer_norm_eps UpperCAmelCase_ : Any = hidden_act UpperCAmelCase_ : Any = qkv_bias @classmethod def A__ ( cls: Tuple ,lowerCamelCase_: Union[str, os.PathLike] ,**lowerCamelCase_: Union[str, Any] ) -> "PretrainedConfig": cls._set_token_in_kwargs(lowerCamelCase_ ) UpperCAmelCase_ : Optional[Any] = cls.get_config_dict(lowerCamelCase_ ,**lowerCamelCase_ ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCAmelCase_ : Union[str, Any] = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls ,"""model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase_ ,**lowerCamelCase_ ) class _snake_case ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' A__ : Union[str, Any] = "blip_2_qformer" def __init__( self: Optional[Any] ,lowerCamelCase_: Optional[Any]=30522 ,lowerCamelCase_: List[Any]=768 ,lowerCamelCase_: Optional[int]=12 ,lowerCamelCase_: Tuple=12 ,lowerCamelCase_: Union[str, Any]=3072 ,lowerCamelCase_: List[str]="gelu" ,lowerCamelCase_: Tuple=0.1 ,lowerCamelCase_: List[Any]=0.1 ,lowerCamelCase_: Any=512 ,lowerCamelCase_: Any=0.0_2 ,lowerCamelCase_: Union[str, Any]=1e-12 ,lowerCamelCase_: Tuple=0 ,lowerCamelCase_: Dict="absolute" ,lowerCamelCase_: List[Any]=2 ,lowerCamelCase_: List[Any]=1408 ,**lowerCamelCase_: str ,) -> Union[str, Any]: super().__init__(pad_token_id=lowerCamelCase_ ,**lowerCamelCase_ ) UpperCAmelCase_ : str = vocab_size UpperCAmelCase_ : Dict = hidden_size UpperCAmelCase_ : str = num_hidden_layers UpperCAmelCase_ : Tuple = num_attention_heads UpperCAmelCase_ : Any = hidden_act UpperCAmelCase_ : Optional[Any] = intermediate_size UpperCAmelCase_ : Dict = hidden_dropout_prob UpperCAmelCase_ : Any = attention_probs_dropout_prob UpperCAmelCase_ : Optional[Any] = max_position_embeddings UpperCAmelCase_ : List[Any] = initializer_range UpperCAmelCase_ : Any = layer_norm_eps UpperCAmelCase_ : Tuple = position_embedding_type UpperCAmelCase_ : Optional[int] = cross_attention_frequency UpperCAmelCase_ : Optional[Any] = encoder_hidden_size @classmethod def A__ ( cls: Dict ,lowerCamelCase_: Union[str, os.PathLike] ,**lowerCamelCase_: Any ) -> "PretrainedConfig": cls._set_token_in_kwargs(lowerCamelCase_ ) UpperCAmelCase_ : Optional[int] = cls.get_config_dict(lowerCamelCase_ ,**lowerCamelCase_ ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCAmelCase_ : Optional[Any] = config_dict["""qformer_config"""] if "model_type" in config_dict and hasattr(cls ,"""model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase_ ,**lowerCamelCase_ ) class _snake_case ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' A__ : str = "blip-2" A__ : Optional[Any] = True def __init__( self: Union[str, Any] ,lowerCamelCase_: Any=None ,lowerCamelCase_: List[str]=None ,lowerCamelCase_: List[Any]=None ,lowerCamelCase_: Dict=32 ,**lowerCamelCase_: List[str] ) -> Union[str, Any]: super().__init__(**lowerCamelCase_ ) if vision_config is None: UpperCAmelCase_ : List[str] = {} logger.info("""vision_config is None. initializing the Blip2VisionConfig with default values.""" ) if qformer_config is None: UpperCAmelCase_ : Any = {} logger.info("""qformer_config is None. Initializing the Blip2QFormerConfig with default values.""" ) if text_config is None: UpperCAmelCase_ : Union[str, Any] = {} logger.info("""text_config is None. Initializing the text config with default values (`OPTConfig`).""" ) UpperCAmelCase_ : Optional[int] = BlipaVisionConfig(**lowerCamelCase_ ) UpperCAmelCase_ : List[str] = BlipaQFormerConfig(**lowerCamelCase_ ) UpperCAmelCase_ : List[str] = text_config["""model_type"""] if """model_type""" in text_config else """opt""" UpperCAmelCase_ : List[str] = CONFIG_MAPPING[text_model_type](**lowerCamelCase_ ) UpperCAmelCase_ : Union[str, Any] = self.text_config.tie_word_embeddings UpperCAmelCase_ : Dict = self.text_config.is_encoder_decoder UpperCAmelCase_ : Union[str, Any] = num_query_tokens UpperCAmelCase_ : List[Any] = self.vision_config.hidden_size UpperCAmelCase_ : int = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES UpperCAmelCase_ : str = 1.0 UpperCAmelCase_ : str = 0.0_2 @classmethod def A__ ( cls: int ,lowerCamelCase_: BlipaVisionConfig ,lowerCamelCase_: BlipaQFormerConfig ,lowerCamelCase_: PretrainedConfig ,**lowerCamelCase_: Any ,) -> List[str]: return cls( vision_config=vision_config.to_dict() ,qformer_config=qformer_config.to_dict() ,text_config=text_config.to_dict() ,**lowerCamelCase_ ,) def A__ ( self: List[str] ) -> int: UpperCAmelCase_ : Union[str, Any] = copy.deepcopy(self.__dict__ ) UpperCAmelCase_ : Union[str, Any] = self.vision_config.to_dict() UpperCAmelCase_ : Optional[Any] = self.qformer_config.to_dict() UpperCAmelCase_ : Dict = self.text_config.to_dict() UpperCAmelCase_ : Optional[int] = self.__class__.model_type return output
350
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed UpperCamelCase_ = '''true''' def lowerCamelCase_ ( _a : List[Any] , _a : List[str]=82 , _a : Tuple=16 ): '''simple docstring''' set_seed(42 ) UpperCAmelCase_ : int = RegressionModel() UpperCAmelCase_ : List[Any] = deepcopy(_a ) UpperCAmelCase_ : Tuple = RegressionDataset(length=_a ) UpperCAmelCase_ : int = DataLoader(_a , batch_size=_a ) model.to(accelerator.device ) UpperCAmelCase_ , UpperCAmelCase_ : Dict = accelerator.prepare(_a , _a ) return model, ddp_model, dataloader def lowerCamelCase_ ( _a : Accelerator , _a : Optional[int]=False ): '''simple docstring''' UpperCAmelCase_ : Any = AutoTokenizer.from_pretrained("""hf-internal-testing/mrpc-bert-base-cased""" ) UpperCAmelCase_ : int = load_dataset("""glue""" , """mrpc""" , split="""validation""" ) def tokenize_function(_a : str ): UpperCAmelCase_ : List[Any] = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=_a , max_length=_a ) return outputs with accelerator.main_process_first(): UpperCAmelCase_ : List[str] = dataset.map( _a , batched=_a , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) UpperCAmelCase_ : Tuple = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(_a : List[str] ): if use_longest: return tokenizer.pad(_a , padding="""longest""" , return_tensors="""pt""" ) return tokenizer.pad(_a , padding="""max_length""" , max_length=128 , return_tensors="""pt""" ) return DataLoader(_a , shuffle=_a , collate_fn=_a , batch_size=16 ) def lowerCamelCase_ ( _a : Any , _a : int ): '''simple docstring''' UpperCAmelCase_ : int = Accelerator(dispatch_batches=_a , split_batches=_a ) UpperCAmelCase_ : Dict = get_dataloader(_a , not dispatch_batches ) UpperCAmelCase_ : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained( """hf-internal-testing/mrpc-bert-base-cased""" , return_dict=_a ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = accelerator.prepare(_a , _a ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def lowerCamelCase_ ( _a : Optional[int] , _a : Optional[Any] , _a : str ): '''simple docstring''' UpperCAmelCase_ : List[str] = [] for batch in dataloader: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = batch.values() with torch.no_grad(): UpperCAmelCase_ : str = model(_a ) UpperCAmelCase_ , UpperCAmelCase_ : Any = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = [], [] for logit, targ in logits_and_targets: logits.append(_a ) targs.append(_a ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = torch.cat(_a ), torch.cat(_a ) return logits, targs def lowerCamelCase_ ( _a : Accelerator , _a : str=82 , _a : str=False , _a : Dict=False , _a : Dict=16 ): '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[str] = get_basic_setup(_a , _a , _a ) UpperCAmelCase_ , UpperCAmelCase_ : Any = generate_predictions(_a , _a , _a ) assert ( len(_a ) == num_samples ), F'''Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_a )}''' def lowerCamelCase_ ( _a : bool = False , _a : bool = False ): '''simple docstring''' UpperCAmelCase_ : List[str] = evaluate.load("""glue""" , """mrpc""" ) UpperCAmelCase_ , UpperCAmelCase_ : str = get_mrpc_setup(_a , _a ) # First do baseline UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = setup["""no"""] model.to(_a ) model.eval() for batch in dataloader: batch.to(_a ) with torch.inference_mode(): UpperCAmelCase_ : str = model(**_a ) UpperCAmelCase_ : Any = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_a , references=batch["""labels"""] ) UpperCAmelCase_ : str = metric.compute() # Then do distributed UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = setup["""ddp"""] model.eval() for batch in dataloader: with torch.inference_mode(): UpperCAmelCase_ : List[str] = model(**_a ) UpperCAmelCase_ : str = outputs.logits.argmax(dim=-1 ) UpperCAmelCase_ : Union[str, Any] = batch["""labels"""] UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_a , references=_a ) UpperCAmelCase_ : str = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), F'''Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n''' def lowerCamelCase_ ( ): '''simple docstring''' UpperCAmelCase_ : Any = Accelerator(split_batches=_a , dispatch_batches=_a ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("""**Testing gather_for_metrics**""" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`''' ) test_mrpc(_a , _a ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("""**Test torch metrics**""" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: UpperCAmelCase_ : Optional[int] = Accelerator(split_batches=_a , dispatch_batches=_a ) if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99''' ) test_torch_metrics(_a , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("""**Test last batch is not dropped when perfectly divisible**""" ) UpperCAmelCase_ : str = Accelerator() test_torch_metrics(_a , 512 ) accelerator.state._reset_state() def lowerCamelCase_ ( _a : Optional[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
59
0
import argparse import copy def UpperCamelCase_( _snake_case : List[str] ): """simple docstring""" __a ={} with open(_snake_case ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: __a =[] _list.append([line.split()[1], line.split()[2]] ) __a =_list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: __a =[] _list.append([line.split()[0], line.split()[2]] ) __a =_list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def UpperCamelCase_( _snake_case : Any , _snake_case : str ): """simple docstring""" with open(_snake_case ) as f: __a =f.read(1 ) __a =start_node __a =[] __a =start_node __a =0 while visiting not in first_solution: __a =10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(_snake_case ) and k[0] not in first_solution: __a =k[1] __a =k[0] first_solution.append(_snake_case ) __a =distance_of_first_solution + int(_snake_case ) __a =best_node first_solution.append(_snake_case ) __a =0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 __a =( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def UpperCamelCase_( _snake_case : List[str] , _snake_case : Dict ): """simple docstring""" __a =[] for n in solution[1:-1]: __a =solution.index(_snake_case ) for kn in solution[1:-1]: __a =solution.index(_snake_case ) if n == kn: continue __a =copy.deepcopy(_snake_case ) __a =kn __a =n __a =0 for k in _tmp[:-1]: __a =_tmp[_tmp.index(_snake_case ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: __a =distance + int(i[1] ) _tmp.append(_snake_case ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) __a =len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda _snake_case : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def UpperCamelCase_( _snake_case : List[str] , _snake_case : Tuple , _snake_case : Any , _snake_case : List[str] , _snake_case : List[Any] ): """simple docstring""" __a =1 __a =first_solution __a =[] __a =distance_of_first_solution __a =solution while count <= iters: __a =find_neighborhood(_snake_case , _snake_case ) __a =0 __a =neighborhood[index_of_best_solution] __a =len(_snake_case ) - 1 __a =False while not found: __a =0 while i < len(_snake_case ): if best_solution[i] != solution[i]: __a =best_solution[i] __a =solution[i] break __a =i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) __a =True __a =best_solution[:-1] __a =neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: __a =cost __a =solution else: __a =index_of_best_solution + 1 __a =neighborhood[index_of_best_solution] if len(_snake_case ) >= size: tabu_list.pop(0 ) __a =count + 1 return best_solution_ever, best_cost def UpperCamelCase_( _snake_case : List[str]=None ): """simple docstring""" __a =generate_neighbours(args.File ) __a , __a =generate_first_solution( args.File , _snake_case ) __a , __a =tabu_search( _snake_case , _snake_case , _snake_case , args.Iterations , args.Size , ) print(F'Best solution: {best_sol}, with total distance: {best_cost}.' ) if __name__ == "__main__": _lowerCAmelCase : Dict = argparse.ArgumentParser(description="Tabu Search") parser.add_argument( "-f", "--File", type=str, help="Path to the file containing the data", required=True, ) parser.add_argument( "-i", "--Iterations", type=int, help="How many iterations the algorithm should perform", required=True, ) parser.add_argument( "-s", "--Size", type=int, help="Size of the tabu list", required=True ) # Pass the arguments to main method main(parser.parse_args())
218
import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase : List[Any] = get_tests_dir("fixtures/test_sentencepiece_with_bytefallback.model") @require_sentencepiece @require_tokenizers class __magic_name__ ( lowerCAmelCase_ , unittest.TestCase ): SCREAMING_SNAKE_CASE = GPTSwaTokenizer SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def __magic_name__ ( self ) -> Dict: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __a =GPTSwaTokenizer(__snake_case , eos_token='<unk>' , bos_token='<unk>' , pad_token='<unk>' ) tokenizer.save_pretrained(self.tmpdirname ) def __magic_name__ ( self , __snake_case ) -> Tuple: '''simple docstring''' __a ='This is a test' __a ='This is a test' return input_text, output_text def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' __a ='<s>' __a =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__snake_case ) , __snake_case ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__snake_case ) , __snake_case ) def __magic_name__ ( self ) -> Dict: '''simple docstring''' __a =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , 'j' ) self.assertEqual(len(__snake_case ) , 2000 ) def __magic_name__ ( self ) -> Optional[Any]: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 2000 ) def __magic_name__ ( self ) -> Any: '''simple docstring''' __a =GPTSwaTokenizer(__snake_case ) __a =tokenizer.tokenize('This is a test' ) self.assertListEqual(__snake_case , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__snake_case ) , [465, 287, 265, 631, 842] ) __a =tokenizer.tokenize('I was born in 92000, and this is falsé.' ) # fmt: off self.assertListEqual( __snake_case , ['▁I', '▁was', '▁bor', 'n', '▁in', '▁', '<0x39>', '2', '0', '0', '0', ',', '▁and', '▁this', '▁is', '▁f', 'al', 's', '<0xC3>', '<0xA9>', '.'] , ) # fmt: on __a =tokenizer.convert_tokens_to_ids(__snake_case ) self.assertListEqual( __snake_case , [262, 272, 1525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __a =tokenizer.convert_ids_to_tokens(__snake_case ) # fmt: off self.assertListEqual( __snake_case , ['▁I', '▁was', '▁bor', 'n', '▁in', '▁', '<0x39>', '2', '0', '0', '0', ',', '▁and', '▁this', '▁is', '▁f', 'al', 's', '<0xC3>', '<0xA9>', '.'] ) # fmt: on def __magic_name__ ( self ) -> Optional[int]: '''simple docstring''' __a =GPTSwaTokenizer(__snake_case ) __a =['This is a test', 'I was born in 92000, and this is falsé.'] __a =[ [465, 287, 265, 631, 842], [262, 272, 1525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(__snake_case , __snake_case ): self.assertListEqual(tokenizer.encode_fast(__snake_case ) , __snake_case ) # Test that decode_fast returns the input text for text, token_ids in zip(__snake_case , __snake_case ): self.assertEqual(tokenizer.decode_fast(__snake_case ) , __snake_case ) @slow def __magic_name__ ( self ) -> List[str]: '''simple docstring''' __a =[ '<|python|>def fibonacci(n)\n if n < 0:\n print(\'Incorrect input\')', 'Hey there, how are you doing this fine day?', 'This is a text with a trailing spaces followed by a dot .', 'Häj sväjs lillebrör! =)', 'Det är inget fel på Mr. Cool', ] # fmt: off __a ={'input_ids': [[6_3423, 5, 6811, 1_4954, 282, 816, 3821, 6_3466, 6_3425, 6_3462, 18, 6_3978, 678, 301, 1320, 6_3423, 6_3455, 6_3458, 18, 6_3982, 4246, 3940, 1901, 4_7789, 5547, 1_8994], [1_9630, 1100, 6_3446, 1342, 633, 544, 4488, 593, 5102, 2416, 6_3495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1652, 428, 268, 1936, 515, 268, 5_8593, 2_2413, 9106, 546, 268, 3_3213, 6_3979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_5130, 6_3450, 924, 6_3449, 2249, 4062, 1558, 318, 6_3504, 2_1498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2827, 2559, 332, 6575, 6_3443, 2_6801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=__snake_case , model_name='AI-Sweden/gpt-sw3-126m' , sequences=__snake_case , )
218
1
def __A ( _lowercase , _lowercase ): '''simple docstring''' _A = '''''' for i in table: res += inp[i - 1] return res def __A ( _lowercase ): '''simple docstring''' return data[1:] + data[0] def __A ( _lowercase , _lowercase ): '''simple docstring''' _A = '''''' for i in range(len(_lowercase ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def __A ( _lowercase , _lowercase ): '''simple docstring''' _A = int('''0b''' + data[0] + data[-1] , 2 ) _A = int('''0b''' + data[1:3] , 2 ) return bin(s[row][col] )[2:] def __A ( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ): '''simple docstring''' _A = message[:4] _A = message[4:] _A = apply_table(_lowercase , _lowercase ) _A = xor(_lowercase , _lowercase ) _A = apply_sbox(_lowercase , temp[:4] ) # noqa: E741 _A = apply_sbox(_lowercase , temp[4:] ) _A = '''0''' * (2 - len(_lowercase )) + l # noqa: E741 _A = '''0''' * (2 - len(_lowercase )) + r _A = apply_table(l + r , _lowercase ) _A = xor(_lowercase , _lowercase ) return temp + right if __name__ == "__main__": __A = input('Enter 10 bit key: ') __A = input('Enter 8 bit message: ') __A = [6, 3, 7, 4, 8, 5, 10, 9] __A = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] __A = [2, 4, 3, 1] __A = [2, 6, 3, 1, 4, 8, 5, 7] __A = [4, 1, 3, 5, 7, 2, 8, 6] __A = [4, 1, 2, 3, 2, 3, 4, 1] __A = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] __A = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation __A = apply_table(key, paa_table) __A = temp[:5] __A = temp[5:] __A = left_shift(left) __A = left_shift(right) __A = apply_table(left + right, pa_table) __A = left_shift(left) __A = left_shift(right) __A = left_shift(left) __A = left_shift(right) __A = apply_table(left + right, pa_table) # encryption __A = apply_table(message, IP) __A = function(expansion, sa, sa, keya, temp) __A = temp[4:] + temp[:4] __A = function(expansion, sa, sa, keya, temp) __A = apply_table(temp, IP_inv) print('Cipher text is:', CT) # decryption __A = apply_table(CT, IP) __A = function(expansion, sa, sa, keya, temp) __A = temp[4:] + temp[:4] __A = function(expansion, sa, sa, keya, temp) __A = apply_table(temp, IP_inv) print('Plain text after decypting is:', PT)
75
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class SCREAMING_SNAKE_CASE ( snake_case ): """simple docstring""" @require_torch def __A ( self: Dict ) -> Optional[int]: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _A = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache _A = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(__A ) BertModel.from_pretrained(__A ) BertTokenizer.from_pretrained(__A ) pipeline(task='''fill-mask''' , model=__A ) # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed _A = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _A = '''1''' _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def __A ( self: Dict ) -> Tuple: # python one-liner segments # this must be loaded before socket.socket is monkey-patched _A = ''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache _A = '''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(__A ) BertModel.from_pretrained(__A ) BertTokenizer.from_pretrained(__A ) pipeline(task='''fill-mask''' , model=__A ) # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed _A = self.get_env() _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def __A ( self: Any ) -> Optional[Any]: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _A = ''' from transformers import BertConfig, BertModel, BertTokenizer ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert-sharded" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print("success") ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled") socket.socket = offline_socket ''' # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed _A = self.get_env() _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # next emulate no network _A = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _A = '''1''' _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def __A ( self: Optional[int] ) -> Dict: _A = ''' from transformers import pipeline ''' _A = ''' mname = "hf-internal-testing/tiny-random-bert" pipe = pipeline(model=mname) ''' _A = ''' import socket def offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled") socket.socket = offline_socket ''' _A = self.get_env() _A = '''1''' _A = [sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( '''You cannot infer task automatically within `pipeline` when using offline mode''' , result.stderr.decode().replace('''\n''' , '''''' ) , ) @require_torch def __A ( self: Optional[int] ) -> int: _A = ''' from transformers import AutoModel ''' _A = ''' mname = "hf-internal-testing/test_dynamic_model" AutoModel.from_pretrained(mname, trust_remote_code=True) print("success") ''' # baseline - just load from_pretrained with normal network _A = [sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed _A = self.get_env() _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _A = '''1''' _A = subprocess.run(__A , env=__A , check=__A , capture_output=__A ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() )
75
1
"""simple docstring""" def _lowerCAmelCase ( UpperCamelCase_ = 50 ): __SCREAMING_SNAKE_CASE = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
100
'''simple docstring''' from maths.prime_factors import prime_factors def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" if not isinstance(UpperCamelCase , UpperCamelCase ): lowerCAmelCase__ : int = f"""Input value of [number={number}] must be an integer""" raise TypeError(UpperCamelCase ) if number < 1: raise ValueError("""Input must be a positive integer""" ) return -1 if len(prime_factors(UpperCamelCase ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
37
0
"""simple docstring""" from __future__ import annotations def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ) -> tuple: if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif electron_conc < 0: raise ValueError("""Electron concentration cannot be negative in a semiconductor""" ) elif hole_conc < 0: raise ValueError("""Hole concentration cannot be negative in a semiconductor""" ) elif intrinsic_conc < 0: raise ValueError( """Intrinsic concentration cannot be negative in a semiconductor""" ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
166
"""simple docstring""" def a__ ( lowerCAmelCase ) -> bool: if not isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase__ : Tuple = F"""Input value of [number={number}] must be an integer""" raise TypeError(lowerCAmelCase ) if number < 0: return False UpperCAmelCase__ : List[str] = number * number while number > 0: if number % 10 != number_square % 10: return False number //= 10 number_square //= 10 return True if __name__ == "__main__": import doctest doctest.testmod()
166
1
'''simple docstring''' import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowercase__ ( __lowercase : Union[str, Any] , __lowercase : str , __lowercase : Optional[Any] ) -> List[Any]: """simple docstring""" if openai_config_file == "": __UpperCamelCase = OpenAIGPTConfig() else: __UpperCamelCase = OpenAIGPTConfig.from_json_file(UpperCamelCase_ ) __UpperCamelCase = OpenAIGPTModel(UpperCamelCase_ ) # Load weights from numpy load_tf_weights_in_openai_gpt(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Save pytorch-model __UpperCamelCase = pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase = pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict() , UpperCamelCase_ ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(UpperCamelCase_ , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": a__ : Any =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--openai_checkpoint_folder_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--openai_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained OpenAI model. \n''' '''This specifies the model architecture.''' ), ) a__ : Tuple =parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
53
"""simple docstring""" import baseaa def _A ( UpperCamelCase_ : str) -> bytes: '''simple docstring''' return baseaa.baaencode(string.encode("utf-8")) def _A ( UpperCamelCase_ : bytes) -> str: '''simple docstring''' return baseaa.baadecode(UpperCamelCase_).decode("utf-8") if __name__ == "__main__": _a = 'Hello World!' _a = baseaa_encode(test) print(encoded) _a = baseaa_decode(encoded) print(decoded)
17
0
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __SCREAMING_SNAKE_CASE ( __lowercase , unittest.TestCase): _SCREAMING_SNAKE_CASE : Union[str, Any] = UnCLIPImageVariationPipeline _SCREAMING_SNAKE_CASE : str = IMAGE_VARIATION_PARAMS - {'''height''', '''width''', '''guidance_scale'''} _SCREAMING_SNAKE_CASE : Tuple = IMAGE_VARIATION_BATCH_PARAMS _SCREAMING_SNAKE_CASE : Tuple = [ '''generator''', '''return_dict''', '''decoder_num_inference_steps''', '''super_res_num_inference_steps''', ] _SCREAMING_SNAKE_CASE : int = False @property def UpperCamelCase__ ( self ): """simple docstring""" return 32 @property def UpperCamelCase__ ( self ): """simple docstring""" return 32 @property def UpperCamelCase__ ( self ): """simple docstring""" return self.time_input_dim @property def UpperCamelCase__ ( self ): """simple docstring""" return self.time_input_dim * 4 @property def UpperCamelCase__ ( self ): """simple docstring""" return 1_00 @property def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModelWithProjection(_UpperCamelCase ) @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(_UpperCamelCase ) @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ = { 'clip_embeddings_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'cross_attention_dim': self.cross_attention_dim, } lowerCAmelCase__ = UnCLIPTextProjModel(**_UpperCamelCase ) return model @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ = { 'sample_size': 32, # RGB in channels 'in_channels': 3, # Out channels is double in channels because predicts mean and variance 'out_channels': 6, 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': 'identity', } lowerCAmelCase__ = UNetaDConditionModel(**_UpperCamelCase ) return model @property def UpperCamelCase__ ( self ): """simple docstring""" return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase__ = UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def UpperCamelCase__ ( self ): """simple docstring""" # seeded differently to get different unet than `self.dummy_super_res_first` torch.manual_seed(1 ) lowerCAmelCase__ = UNetaDModel(**self.dummy_super_res_kwargs ) return model def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.dummy_decoder lowerCAmelCase__ = self.dummy_text_proj lowerCAmelCase__ = self.dummy_text_encoder lowerCAmelCase__ = self.dummy_tokenizer lowerCAmelCase__ = self.dummy_super_res_first lowerCAmelCase__ = self.dummy_super_res_last lowerCAmelCase__ = UnCLIPScheduler( variance_type='learned_range' , prediction_type='epsilon' , num_train_timesteps=10_00 , ) lowerCAmelCase__ = UnCLIPScheduler( variance_type='fixed_small_log' , prediction_type='epsilon' , num_train_timesteps=10_00 , ) lowerCAmelCase__ = CLIPImageProcessor(crop_size=32 , size=32 ) lowerCAmelCase__ = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase=0 , _UpperCamelCase=True ): """simple docstring""" lowerCAmelCase__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(_UpperCamelCase ) ).to(_UpperCamelCase ) if str(_UpperCamelCase ).startswith('mps' ): lowerCAmelCase__ = torch.manual_seed(_UpperCamelCase ) else: lowerCAmelCase__ = torch.Generator(device=_UpperCamelCase ).manual_seed(_UpperCamelCase ) if pil_image: lowerCAmelCase__ = input_image * 0.5 + 0.5 lowerCAmelCase__ = input_image.clamp(0 , 1 ) lowerCAmelCase__ = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() lowerCAmelCase__ = DiffusionPipeline.numpy_to_pil(_UpperCamelCase )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = 'cpu' lowerCAmelCase__ = self.get_dummy_components() lowerCAmelCase__ = self.pipeline_class(**_UpperCamelCase ) lowerCAmelCase__ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) lowerCAmelCase__ = pipe(**_UpperCamelCase ) lowerCAmelCase__ = output.images lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) lowerCAmelCase__ = pipe( **_UpperCamelCase , return_dict=_UpperCamelCase , )[0] lowerCAmelCase__ = image[0, -3:, -3:, -1] lowerCAmelCase__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ = np.array( [ 0.99_97, 0.00_02, 0.99_97, 0.99_97, 0.99_69, 0.00_23, 0.99_97, 0.99_69, 0.99_70, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = 'cpu' lowerCAmelCase__ = self.get_dummy_components() lowerCAmelCase__ = self.pipeline_class(**_UpperCamelCase ) lowerCAmelCase__ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) lowerCAmelCase__ = pipe(**_UpperCamelCase ) lowerCAmelCase__ = output.images lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) lowerCAmelCase__ = pipe( **_UpperCamelCase , return_dict=_UpperCamelCase , )[0] lowerCAmelCase__ = image[0, -3:, -3:, -1] lowerCAmelCase__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCAmelCase__ = np.array([0.99_97, 0.00_03, 0.99_97, 0.99_97, 0.99_70, 0.00_24, 0.99_97, 0.99_71, 0.99_71] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = 'cpu' lowerCAmelCase__ = self.get_dummy_components() lowerCAmelCase__ = self.pipeline_class(**_UpperCamelCase ) lowerCAmelCase__ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) lowerCAmelCase__ = [ pipeline_inputs['image'], pipeline_inputs['image'], ] lowerCAmelCase__ = pipe(**_UpperCamelCase ) lowerCAmelCase__ = output.images lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) lowerCAmelCase__ = [ tuple_pipeline_inputs['image'], tuple_pipeline_inputs['image'], ] lowerCAmelCase__ = pipe( **_UpperCamelCase , return_dict=_UpperCamelCase , )[0] lowerCAmelCase__ = image[0, -3:, -3:, -1] lowerCAmelCase__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) lowerCAmelCase__ = np.array( [ 0.99_97, 0.99_89, 0.00_08, 0.00_21, 0.99_60, 0.00_18, 0.00_14, 0.00_02, 0.99_33, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = torch.device('cpu' ) class __SCREAMING_SNAKE_CASE : _SCREAMING_SNAKE_CASE : Tuple = 1 lowerCAmelCase__ = self.get_dummy_components() lowerCAmelCase__ = self.pipeline_class(**_UpperCamelCase ) lowerCAmelCase__ = pipe.to(_UpperCamelCase ) pipe.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase__ = torch.Generator(device=_UpperCamelCase ).manual_seed(0 ) lowerCAmelCase__ = pipe.decoder.dtype lowerCAmelCase__ = 1 lowerCAmelCase__ = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) lowerCAmelCase__ = pipe.prepare_latents( _UpperCamelCase , dtype=_UpperCamelCase , device=_UpperCamelCase , generator=_UpperCamelCase , latents=_UpperCamelCase , scheduler=DummyScheduler() ) lowerCAmelCase__ = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) lowerCAmelCase__ = pipe.prepare_latents( _UpperCamelCase , dtype=_UpperCamelCase , device=_UpperCamelCase , generator=_UpperCamelCase , latents=_UpperCamelCase , scheduler=DummyScheduler() ) lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) lowerCAmelCase__ = pipe( **_UpperCamelCase , decoder_latents=_UpperCamelCase , super_res_latents=_UpperCamelCase ).images lowerCAmelCase__ = self.get_dummy_inputs(_UpperCamelCase , pil_image=_UpperCamelCase ) # Don't pass image, instead pass embedding lowerCAmelCase__ = pipeline_inputs.pop('image' ) lowerCAmelCase__ = pipe.image_encoder(_UpperCamelCase ).image_embeds lowerCAmelCase__ = pipe( **_UpperCamelCase , decoder_latents=_UpperCamelCase , super_res_latents=_UpperCamelCase , image_embeddings=_UpperCamelCase , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1E-4 @skip_mps def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = torch_device == 'cpu' # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor lowerCAmelCase__ = 1E-2 self._test_attention_slicing_forward_pass( test_max_difference=_UpperCamelCase , expected_max_diff=_UpperCamelCase ) @skip_mps def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = torch_device == 'cpu' lowerCAmelCase__ = True lowerCAmelCase__ = [ 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] self._test_inference_batch_single_identical( test_max_difference=_UpperCamelCase , relax_max_difference=_UpperCamelCase , additional_params_copy_to_batched_inputs=_UpperCamelCase , ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = [ 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes lowerCAmelCase__ = [2, 3] self._test_inference_batch_consistent( batch_sizes=_UpperCamelCase , additional_params_copy_to_batched_inputs=_UpperCamelCase , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=_UpperCamelCase ) @skip_mps def UpperCamelCase__ ( self ): """simple docstring""" return super().test_dict_tuple_outputs_equivalent() @skip_mps def UpperCamelCase__ ( self ): """simple docstring""" return super().test_save_load_local() @skip_mps def UpperCamelCase__ ( self ): """simple docstring""" return super().test_save_load_optional_components() @slow @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase): def UpperCamelCase__ ( self ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png' ) lowerCAmelCase__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/unclip/karlo_v1_alpha_cat_variation_fp16.npy' ) lowerCAmelCase__ = UnCLIPImageVariationPipeline.from_pretrained( 'kakaobrain/karlo-v1-alpha-image-variations' , torch_dtype=torch.floataa ) lowerCAmelCase__ = pipeline.to(_UpperCamelCase ) pipeline.set_progress_bar_config(disable=_UpperCamelCase ) lowerCAmelCase__ = torch.Generator(device='cpu' ).manual_seed(0 ) lowerCAmelCase__ = pipeline( _UpperCamelCase , generator=_UpperCamelCase , output_type='np' , ) lowerCAmelCase__ = output.images[0] assert image.shape == (2_56, 2_56, 3) assert_mean_pixel_difference(_UpperCamelCase , _UpperCamelCase , 15 )
122
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __snake_case : Any = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __lowercase): _SCREAMING_SNAKE_CASE : List[Any] = ['''pixel_values'''] def __init__( self , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = PILImageResampling.BILINEAR , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = True , _UpperCamelCase = 1 / 2_55 , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" super().__init__(**_UpperCamelCase ) lowerCAmelCase__ = size if size is not None else {'shortest_edge': 2_56} lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) lowerCAmelCase__ = crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} lowerCAmelCase__ = get_size_dict(_UpperCamelCase ) lowerCAmelCase__ = do_resize lowerCAmelCase__ = size lowerCAmelCase__ = resample lowerCAmelCase__ = do_center_crop lowerCAmelCase__ = crop_size lowerCAmelCase__ = do_rescale lowerCAmelCase__ = rescale_factor lowerCAmelCase__ = do_normalize lowerCAmelCase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCAmelCase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = PILImageResampling.BICUBIC , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) lowerCAmelCase__ = get_resize_output_image_size(_UpperCamelCase , size=size['shortest_edge'] , default_to_square=_UpperCamelCase ) return resize(_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = get_size_dict(_UpperCamelCase ) return center_crop(_UpperCamelCase , size=(size['height'], size['width']) , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase ): """simple docstring""" return rescale(_UpperCamelCase , scale=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" return normalize(_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = ChannelDimension.FIRST , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = do_resize if do_resize is not None else self.do_resize lowerCAmelCase__ = size if size is not None else self.size lowerCAmelCase__ = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) lowerCAmelCase__ = resample if resample is not None else self.resample lowerCAmelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop lowerCAmelCase__ = crop_size if crop_size is not None else self.crop_size lowerCAmelCase__ = get_size_dict(_UpperCamelCase ) lowerCAmelCase__ = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase__ = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase__ = image_mean if image_mean is not None else self.image_mean lowerCAmelCase__ = image_std if image_std is not None else self.image_std lowerCAmelCase__ = make_list_of_images(_UpperCamelCase ) if not valid_images(_UpperCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. lowerCAmelCase__ = [to_numpy_array(_UpperCamelCase ) for image in images] if do_resize: lowerCAmelCase__ = [self.resize(image=_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase ) for image in images] if do_center_crop: lowerCAmelCase__ = [self.center_crop(image=_UpperCamelCase , size=_UpperCamelCase ) for image in images] if do_rescale: lowerCAmelCase__ = [self.rescale(image=_UpperCamelCase , scale=_UpperCamelCase ) for image in images] if do_normalize: lowerCAmelCase__ = [self.normalize(image=_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase ) for image in images] lowerCAmelCase__ = [to_channel_dimension_format(_UpperCamelCase , _UpperCamelCase ) for image in images] lowerCAmelCase__ = {'pixel_values': images} return BatchFeature(data=_UpperCamelCase , tensor_type=_UpperCamelCase )
122
1
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class UpperCamelCase_ ( __magic_name__ ): def _lowercase( self , A ) -> float: return 0.0 def __lowerCamelCase ( _lowercase , _lowercase ) -> tuple[int | float, int | float]: UpperCAmelCase : Dict = min([-2_0, np.min(fft_results[1 : samplerate // 2 - 1] )] ) UpperCAmelCase : Any = max([2_0, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def __lowerCamelCase ( _lowercase , _lowercase ) -> None: UpperCAmelCase : int = 5_1_2 UpperCAmelCase : List[str] = [1] + [0] * (size - 1) UpperCAmelCase : str = [filter_type.process(_lowercase ) for item in inputs] UpperCAmelCase : Any = [0] * (samplerate - size) # zero-padding outputs += filler UpperCAmelCase : Optional[Any] = np.abs(np.fft.fft(_lowercase ) ) UpperCAmelCase : Optional[int] = 2_0 * np.logaa(_lowercase ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(2_4 , samplerate / 2 - 1 ) plt.xlabel("""Frequency (Hz)""" ) plt.xscale("""log""" ) # Display within reasonable bounds UpperCAmelCase : Any = get_bounds(_lowercase , _lowercase ) plt.ylim(max([-8_0, bounds[0]] ) , min([8_0, bounds[1]] ) ) plt.ylabel("""Gain (dB)""" ) plt.plot(_lowercase ) plt.show() def __lowerCamelCase ( _lowercase , _lowercase ) -> None: UpperCAmelCase : Optional[Any] = 5_1_2 UpperCAmelCase : Union[str, Any] = [1] + [0] * (size - 1) UpperCAmelCase : Optional[int] = [filter_type.process(_lowercase ) for item in inputs] UpperCAmelCase : List[Any] = [0] * (samplerate - size) # zero-padding outputs += filler UpperCAmelCase : Optional[Any] = np.angle(np.fft.fft(_lowercase ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(2_4 , samplerate / 2 - 1 ) plt.xlabel("""Frequency (Hz)""" ) plt.xscale("""log""" ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel("""Phase shift (Radians)""" ) plt.plot(np.unwrap(_lowercase , -2 * pi ) ) plt.show()
265
'''simple docstring''' import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __lowerCamelCase ( ) -> Any: raise RuntimeError("""CUDA out of memory.""" ) class UpperCamelCase_ ( nn.Module ): def __init__( self ) -> Any: super().__init__() UpperCAmelCase : Tuple = nn.Linear(3 , 4 ) UpperCAmelCase : Tuple = nn.BatchNormad(4 ) UpperCAmelCase : int = nn.Linear(4 , 5 ) def _lowercase( self , A ) -> Any: return self.lineara(self.batchnorm(self.lineara(A ) ) ) class UpperCamelCase_ ( unittest.TestCase ): def _lowercase( self ) -> Optional[Any]: UpperCAmelCase : Optional[int] = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(A ): nonlocal batch_sizes batch_sizes.append(A ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(A , [128, 64, 32, 16, 8] ) def _lowercase( self ) -> Any: UpperCAmelCase : Optional[Any] = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(A , A ): nonlocal batch_sizes batch_sizes.append(A ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga UpperCAmelCase , UpperCAmelCase : Optional[int] = mock_training_loop_function("""hello""" ) self.assertListEqual(A , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, """hello"""] ) def _lowercase( self ) -> Any: @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(A ): pass with self.assertRaises(A ) as cm: mock_training_loop_function() self.assertIn("""No executable batch size found, reached zero.""" , cm.exception.args[0] ) def _lowercase( self ) -> Optional[int]: @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(A ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(A ) as cm: mock_training_loop_function() self.assertIn("""No executable batch size found, reached zero.""" , cm.exception.args[0] ) def _lowercase( self ) -> Optional[Any]: @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(A , A , A ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(A ) as cm: mock_training_loop_function(128 , """hello""" , """world""" ) self.assertIn("""Batch size was passed into `f`""" , cm.exception.args[0] ) self.assertIn("""`f(arg1='hello', arg2='world')""" , cm.exception.args[0] ) def _lowercase( self ) -> int: @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(A ): raise ValueError("""Oops, we had an error!""" ) with self.assertRaises(A ) as cm: mock_training_loop_function() self.assertIn("""Oops, we had an error!""" , cm.exception.args[0] ) @require_cuda def _lowercase( self ) -> Optional[int]: UpperCAmelCase : Optional[Any] = torch.cuda.memory_allocated() UpperCAmelCase : List[str] = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , A ) UpperCAmelCase : Tuple = release_memory(A ) self.assertEqual(torch.cuda.memory_allocated() , A )
265
1
"""simple docstring""" import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCamelCase_ (__A , unittest.TestCase ): __magic_name__ = AudioLDMPipeline __magic_name__ = TEXT_TO_AUDIO_PARAMS __magic_name__ = TEXT_TO_AUDIO_BATCH_PARAMS __magic_name__ = frozenset( [ '''num_inference_steps''', '''num_waveforms_per_prompt''', '''generator''', '''latents''', '''output_type''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> List[Any]: torch.manual_seed(0 ) UpperCAmelCase_ : Any = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=(32, 64) , class_embed_type="simple_projection" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=lowerCAmelCase_ , ) UpperCAmelCase_ : Any = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" , clip_sample=lowerCAmelCase_ , set_alpha_to_one=lowerCAmelCase_ , ) torch.manual_seed(0 ) UpperCAmelCase_ : Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) UpperCAmelCase_ : Tuple = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , projection_dim=32 , ) UpperCAmelCase_ : Tuple = ClapTextModelWithProjection(lowerCAmelCase_ ) UpperCAmelCase_ : int = RobertaTokenizer.from_pretrained("hf-internal-testing/tiny-random-roberta" , model_max_length=77 ) UpperCAmelCase_ : Tuple = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=16_000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=lowerCAmelCase_ , ) UpperCAmelCase_ : int = SpeechTaHifiGan(lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "vocoder": vocoder, } return components def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Tuple=0 ) -> Tuple: if str(lowerCAmelCase_ ).startswith("mps" ): UpperCAmelCase_ : int = torch.manual_seed(lowerCAmelCase_ ) else: UpperCAmelCase_ : Any = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = { "prompt": "A hammer hitting a wooden surface", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, } return inputs def _SCREAMING_SNAKE_CASE ( self : Any ) -> Union[str, Any]: UpperCAmelCase_ : int = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : List[str] = self.get_dummy_components() UpperCAmelCase_ : List[Any] = AudioLDMPipeline(**lowerCAmelCase_ ) UpperCAmelCase_ : Dict = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = audioldm_pipe(**lowerCAmelCase_ ) UpperCAmelCase_ : Tuple = output.audios[0] assert audio.ndim == 1 assert len(lowerCAmelCase_ ) == 256 UpperCAmelCase_ : Optional[Any] = audio[:10] UpperCAmelCase_ : Any = np.array( [-0.0_0_5_0, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_3, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_3] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Tuple: UpperCAmelCase_ : Tuple = self.get_dummy_components() UpperCAmelCase_ : Optional[int] = AudioLDMPipeline(**lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = audioldm_pipe.to(lowerCAmelCase_ ) UpperCAmelCase_ : int = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = 3 * [inputs["prompt"]] # forward UpperCAmelCase_ : Optional[int] = audioldm_pipe(**lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = output.audios[0] UpperCAmelCase_ : Any = self.get_dummy_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : str = 3 * [inputs.pop("prompt" )] UpperCAmelCase_ : Tuple = audioldm_pipe.tokenizer( lowerCAmelCase_ , padding="max_length" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCAmelCase_ , return_tensors="pt" , ) UpperCAmelCase_ : List[str] = text_inputs["input_ids"].to(lowerCAmelCase_ ) UpperCAmelCase_ : Dict = audioldm_pipe.text_encoder( lowerCAmelCase_ , ) UpperCAmelCase_ : Optional[Any] = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state UpperCAmelCase_ : Tuple = F.normalize(lowerCAmelCase_ , dim=-1 ) UpperCAmelCase_ : int = prompt_embeds # forward UpperCAmelCase_ : Union[str, Any] = audioldm_pipe(**lowerCAmelCase_ ) UpperCAmelCase_ : Dict = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> int: UpperCAmelCase_ : str = self.get_dummy_components() UpperCAmelCase_ : Optional[int] = AudioLDMPipeline(**lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = audioldm_pipe.to(lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : Any = 3 * ["this is a negative prompt"] UpperCAmelCase_ : Union[str, Any] = negative_prompt UpperCAmelCase_ : Dict = 3 * [inputs["prompt"]] # forward UpperCAmelCase_ : List[Any] = audioldm_pipe(**lowerCAmelCase_ ) UpperCAmelCase_ : Optional[int] = output.audios[0] UpperCAmelCase_ : Dict = self.get_dummy_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = 3 * [inputs.pop("prompt" )] UpperCAmelCase_ : List[str] = [] for p in [prompt, negative_prompt]: UpperCAmelCase_ : int = audioldm_pipe.tokenizer( lowerCAmelCase_ , padding="max_length" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCAmelCase_ , return_tensors="pt" , ) UpperCAmelCase_ : List[Any] = text_inputs["input_ids"].to(lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = audioldm_pipe.text_encoder( lowerCAmelCase_ , ) UpperCAmelCase_ : Union[str, Any] = text_embeds.text_embeds # additional L_2 normalization over each hidden-state UpperCAmelCase_ : List[str] = F.normalize(lowerCAmelCase_ , dim=-1 ) embeds.append(lowerCAmelCase_ ) UpperCAmelCase_ : Tuple = embeds # forward UpperCAmelCase_ : str = audioldm_pipe(**lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : str ) -> Dict: UpperCAmelCase_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : str = self.get_dummy_components() UpperCAmelCase_ : List[Any] = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = AudioLDMPipeline(**lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = self.get_dummy_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : Optional[int] = "egg cracking" UpperCAmelCase_ : str = audioldm_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = output.audios[0] assert audio.ndim == 1 assert len(lowerCAmelCase_ ) == 256 UpperCAmelCase_ : List[str] = audio[:10] UpperCAmelCase_ : str = np.array( [-0.0_0_5_1, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_4, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_2] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : str ) -> str: UpperCAmelCase_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : List[Any] = self.get_dummy_components() UpperCAmelCase_ : Optional[int] = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) UpperCAmelCase_ : Dict = AudioLDMPipeline(**lowerCAmelCase_ ) UpperCAmelCase_ : Optional[int] = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : Tuple = "A hammer hitting a wooden surface" # test num_waveforms_per_prompt=1 (default) UpperCAmelCase_ : Dict = audioldm_pipe(lowerCAmelCase_ , num_inference_steps=2 ).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts UpperCAmelCase_ : List[Any] = 2 UpperCAmelCase_ : List[str] = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt UpperCAmelCase_ : List[Any] = 2 UpperCAmelCase_ : Optional[Any] = audioldm_pipe(lowerCAmelCase_ , num_inference_steps=2 , num_waveforms_per_prompt=lowerCAmelCase_ ).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts UpperCAmelCase_ : Union[str, Any] = 2 UpperCAmelCase_ : Optional[int] = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=lowerCAmelCase_ ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Tuple: UpperCAmelCase_ : str = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : List[str] = self.get_dummy_components() UpperCAmelCase_ : List[str] = AudioLDMPipeline(**lowerCAmelCase_ ) UpperCAmelCase_ : str = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = audioldm_pipe.vocoder.config.sampling_rate UpperCAmelCase_ : str = self.get_dummy_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : Any = audioldm_pipe(audio_length_in_s=0.0_1_6 , **lowerCAmelCase_ ) UpperCAmelCase_ : str = output.audios[0] assert audio.ndim == 1 assert len(lowerCAmelCase_ ) / vocoder_sampling_rate == 0.0_1_6 UpperCAmelCase_ : Dict = audioldm_pipe(audio_length_in_s=0.0_3_2 , **lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = output.audios[0] assert audio.ndim == 1 assert len(lowerCAmelCase_ ) / vocoder_sampling_rate == 0.0_3_2 def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Any: UpperCAmelCase_ : Tuple = self.get_dummy_components() UpperCAmelCase_ : Any = AudioLDMPipeline(**lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : Tuple = ["hey"] UpperCAmelCase_ : int = audioldm_pipe(lowerCAmelCase_ , num_inference_steps=1 ) UpperCAmelCase_ : Dict = output.audios.shape assert audio_shape == (1, 256) UpperCAmelCase_ : Optional[int] = audioldm_pipe.vocoder.config config.model_in_dim *= 2 UpperCAmelCase_ : str = SpeechTaHifiGan(lowerCAmelCase_ ).to(lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = audioldm_pipe(lowerCAmelCase_ , num_inference_steps=1 ) UpperCAmelCase_ : List[Any] = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[Any]: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : str ) -> int: self._test_inference_batch_single_identical(test_mean_pixel_difference=lowerCAmelCase_ ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _SCREAMING_SNAKE_CASE ( self : str ) -> Any: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase_ ) @slow class UpperCamelCase_ (unittest.TestCase ): def _SCREAMING_SNAKE_CASE ( self : str ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int]="cpu" , lowerCAmelCase_ : List[str]=torch.floataa , lowerCAmelCase_ : str=0 ) -> Optional[int]: UpperCAmelCase_ : Tuple = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) UpperCAmelCase_ : str = np.random.RandomState(lowerCAmelCase_ ).standard_normal((1, 8, 128, 16) ) UpperCAmelCase_ : Any = torch.from_numpy(lowerCAmelCase_ ).to(device=lowerCAmelCase_ , dtype=lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = { "prompt": "A hammer hitting a wooden surface", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 2.5, } return inputs def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> str: UpperCAmelCase_ : Union[str, Any] = AudioLDMPipeline.from_pretrained("cvssp/audioldm" ) UpperCAmelCase_ : int = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : str = self.get_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : Dict = 25 UpperCAmelCase_ : List[Any] = audioldm_pipe(**lowerCAmelCase_ ).audios[0] assert audio.ndim == 1 assert len(lowerCAmelCase_ ) == 81_920 UpperCAmelCase_ : List[str] = audio[77_230:77_240] UpperCAmelCase_ : List[str] = np.array( [-0.4_8_8_4, -0.4_6_0_7, 0.0_0_2_3, 0.5_0_0_7, 0.5_8_9_6, 0.5_1_5_1, 0.3_8_1_3, -0.0_2_0_8, -0.3_6_8_7, -0.4_3_1_5] ) UpperCAmelCase_ : Union[str, Any] = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> List[str]: UpperCAmelCase_ : Optional[Any] = AudioLDMPipeline.from_pretrained("cvssp/audioldm" ) UpperCAmelCase_ : Union[str, Any] = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) UpperCAmelCase_ : Any = audioldm_pipe.to(lowerCAmelCase_ ) audioldm_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) UpperCAmelCase_ : Dict = self.get_inputs(lowerCAmelCase_ ) UpperCAmelCase_ : str = audioldm_pipe(**lowerCAmelCase_ ).audios[0] assert audio.ndim == 1 assert len(lowerCAmelCase_ ) == 81_920 UpperCAmelCase_ : Dict = audio[27_780:27_790] UpperCAmelCase_ : List[Any] = np.array([-0.2_1_3_1, -0.0_8_7_3, -0.0_1_2_4, -0.0_1_8_9, 0.0_5_6_9, 0.1_3_7_3, 0.1_8_8_3, 0.2_8_8_6, 0.3_2_9_7, 0.2_2_1_2] ) UpperCAmelCase_ : int = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
365
"""simple docstring""" from math import factorial def snake_case ( A__ = 1_00 ): return sum(int(A__ ) for x in str(factorial(A__ ) ) ) if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
253
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A__ : List[Any] =logging.get_logger(__name__) A__ : Any =torch.device('''cpu''') def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" _lowerCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = dct.pop(lowerCAmelCase ) _lowerCAmelCase = val def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [] for k in state_dict.keys(): _lowerCAmelCase = k if ".pwconv" in k: _lowerCAmelCase = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _lowerCAmelCase = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _lowerCAmelCase = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _lowerCAmelCase = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _lowerCAmelCase = k_new.split(""".""" ) if ls[2].isdigit(): _lowerCAmelCase = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _lowerCAmelCase = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _lowerCAmelCase = 10_00 _lowerCAmelCase = """huggingface/label-files""" _lowerCAmelCase = """imagenet-1k-id2label.json""" _lowerCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) _lowerCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} _lowerCAmelCase = idalabel _lowerCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _lowerCAmelCase = [3, 3, 6, 4] _lowerCAmelCase = [48, 56, 1_12, 2_20] elif swiftformer_name == "swiftformer_s": _lowerCAmelCase = [3, 3, 9, 6] _lowerCAmelCase = [48, 64, 1_68, 2_24] elif swiftformer_name == "swiftformer_l1": _lowerCAmelCase = [4, 3, 10, 5] _lowerCAmelCase = [48, 96, 1_92, 3_84] elif swiftformer_name == "swiftformer_l3": _lowerCAmelCase = [4, 4, 12, 6] _lowerCAmelCase = [64, 1_28, 3_20, 5_12] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _lowerCAmelCase = torch.hub.load_state_dict_from_url(lowerCAmelCase , map_location="""cpu""" , check_hash=lowerCAmelCase ) else: _lowerCAmelCase = torch.load(lowerCAmelCase , map_location="""cpu""" ) _lowerCAmelCase = checkpoint _lowerCAmelCase = create_rename_keys(lowerCAmelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model _lowerCAmelCase = SwiftFormerForImageClassification(lowerCAmelCase ).eval() hf_model.load_state_dict(lowerCAmelCase ) # prepare test inputs _lowerCAmelCase = prepare_img() _lowerCAmelCase = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _lowerCAmelCase = processor(images=lowerCAmelCase , return_tensors="""pt""" ) # compare outputs from both models _lowerCAmelCase = get_expected_output(lowerCAmelCase ) _lowerCAmelCase = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 10_00] ) assert torch.allclose(hf_logits[0, 0:5] , lowerCAmelCase , atol=1e-3 ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(f"Saving model {swiftformer_name} to {pytorch_dump_folder_path}" ) hf_model.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": A__ : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swiftformer_name''', default='''swiftformer_xs''', choices=['''swiftformer_xs''', '''swiftformer_s''', '''swiftformer_l1''', '''swiftformer_l3'''], type=str, help='''Name of the SwiftFormer model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''./converted_outputs/''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--original_ckpt''', default=None, type=str, help='''Path to the original model checkpoint.''') A__ : Tuple =parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
70
import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class __lowerCamelCase (_a ): _lowercase = 0 _lowercase = False _lowercase = 3.0 class __lowerCamelCase (unittest.TestCase ): def snake_case_ ( self: Any ): '''simple docstring''' self.assertDictEqual(MockClass().to_kwargs(),{} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs(),{'a': 2} ) self.assertDictEqual(MockClass(a=2,b=A_ ).to_kwargs(),{'a': 2, 'b': True} ) self.assertDictEqual(MockClass(a=2,c=2.2_5 ).to_kwargs(),{'a': 2, 'c': 2.2_5} ) @require_cuda def snake_case_ ( self: Optional[int] ): '''simple docstring''' __UpperCamelCase = GradScalerKwargs(init_scale=1024,growth_factor=2 ) AcceleratorState._reset_state() __UpperCamelCase = Accelerator(mixed_precision='fp16',kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) __UpperCamelCase = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale,1_0_2_4.0 ) self.assertEqual(scaler._growth_factor,2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor,0.5 ) self.assertEqual(scaler._growth_interval,2000 ) self.assertEqual(scaler._enabled,A_ ) @require_multi_gpu def snake_case_ ( self: str ): '''simple docstring''' __UpperCamelCase = ['torchrun', F'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] execute_subprocess_async(A_,env=os.environ.copy() ) if __name__ == "__main__": __snake_case = DistributedDataParallelKwargs(bucket_cap_mb=1_5, find_unused_parameters=True) __snake_case = Accelerator(kwargs_handlers=[ddp_scaler]) __snake_case = torch.nn.Linear(1_0_0, 2_0_0) __snake_case = accelerator.prepare(model) # Check the values changed in kwargs __snake_case = '''''' __snake_case = model.bucket_bytes_cap // (1_0_2_4 * 1_0_2_4) if observed_bucket_cap_map != 1_5: error_msg += f"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += f"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += f"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += f"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += f"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
310
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { "BAAI/AltCLIP": "https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json", # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" snake_case_ = '''altclip_text_model''' def __init__( self , lowerCamelCase__=250_002 , lowerCamelCase__=1_024 , lowerCamelCase__=24 , lowerCamelCase__=16 , lowerCamelCase__=4_096 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=514 , lowerCamelCase__=1 , lowerCamelCase__=0.02 , lowerCamelCase__=0.02 , lowerCamelCase__=1e-05 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , lowerCamelCase__="absolute" , lowerCamelCase__=True , lowerCamelCase__=768 , **lowerCamelCase__ , ) -> int: '''simple docstring''' super().__init__(pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = initializer_factor __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = project_dim class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" snake_case_ = '''altclip_vision_model''' def __init__( self , lowerCamelCase__=768 , lowerCamelCase__=3_072 , lowerCamelCase__=512 , lowerCamelCase__=12 , lowerCamelCase__=12 , lowerCamelCase__=3 , lowerCamelCase__=224 , lowerCamelCase__=32 , lowerCamelCase__="quick_gelu" , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=0.02 , lowerCamelCase__=1.0 , **lowerCamelCase__ , ) -> Dict: '''simple docstring''' super().__init__(**lowerCamelCase__ ) __lowerCamelCase = hidden_size __lowerCamelCase = intermediate_size __lowerCamelCase = projection_dim __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = num_channels __lowerCamelCase = patch_size __lowerCamelCase = image_size __lowerCamelCase = initializer_range __lowerCamelCase = initializer_factor __lowerCamelCase = attention_dropout __lowerCamelCase = layer_norm_eps __lowerCamelCase = hidden_act @classmethod def lowercase_ ( cls , lowerCamelCase__ , **lowerCamelCase__ ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get('model_type' ) == "altclip": __lowerCamelCase = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" snake_case_ = '''altclip''' snake_case_ = True def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=768 , lowerCamelCase__=2.65_92 , **lowerCamelCase__ ) -> List[str]: '''simple docstring''' # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). __lowerCamelCase = kwargs.pop('text_config_dict' , lowerCamelCase__ ) __lowerCamelCase = kwargs.pop('vision_config_dict' , lowerCamelCase__ ) super().__init__(**lowerCamelCase__ ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: __lowerCamelCase = {} # This is the complete result when using `text_config_dict`. __lowerCamelCase = AltCLIPTextConfig(**lowerCamelCase__ ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: __lowerCamelCase = ( f"""`{key}` is found in both `text_config_dict` and `text_config` but with different values. """ f"""The value `text_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: __lowerCamelCase = ( f"""`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The """ f"""value `text_config[\"{key}\"]` will be overriden.""" ) logger.warning(lowerCamelCase__ ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: __lowerCamelCase = {} # This is the complete result when using `vision_config_dict`. __lowerCamelCase = AltCLIPVisionConfig(**lowerCamelCase__ ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: __lowerCamelCase = { str(lowerCamelCase__ ): value for key, value in _vision_config_dict['id2label'].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: __lowerCamelCase = ( f"""`{key}` is found in both `vision_config_dict` and `vision_config` but with different """ f"""values. The value `vision_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: __lowerCamelCase = ( f"""`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. """ f"""The value `vision_config[\"{key}\"]` will be overriden.""" ) logger.warning(lowerCamelCase__ ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: __lowerCamelCase = {} logger.info('`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values.' ) if vision_config is None: __lowerCamelCase = {} logger.info('`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values.' ) __lowerCamelCase = AltCLIPTextConfig(**lowerCamelCase__ ) __lowerCamelCase = AltCLIPVisionConfig(**lowerCamelCase__ ) __lowerCamelCase = projection_dim __lowerCamelCase = logit_scale_init_value __lowerCamelCase = 1.0 @classmethod def lowercase_ ( cls , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) -> Optional[Any]: '''simple docstring''' return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **lowerCamelCase__ ) def lowercase_ ( self ) -> Dict: '''simple docstring''' __lowerCamelCase = copy.deepcopy(self.__dict__ ) __lowerCamelCase = self.text_config.to_dict() __lowerCamelCase = self.vision_config.to_dict() __lowerCamelCase = self.__class__.model_type return output
348
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings from diffusers.utils import load_numpy, slow, torch_device from diffusers.utils.testing_utils import require_torch_gpu __A = False class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowercase_ ( self ) -> int: '''simple docstring''' return 12 @property def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' return 12 @property def lowercase_ ( self ) -> List[str]: '''simple docstring''' return 32 @property def lowercase_ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , num_vq_embeddings=self.num_embed , vq_embed_dim=3 , ) return model @property def lowercase_ ( self ) -> List[Any]: '''simple docstring''' __lowerCamelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) return CLIPTextModel(lowerCamelCase__ ) @property def lowercase_ ( self ) -> str: '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = 12 __lowerCamelCase = 12 __lowerCamelCase = { 'attention_bias': True, 'cross_attention_dim': 32, 'attention_head_dim': height * width, 'num_attention_heads': 1, 'num_vector_embeds': self.num_embed, 'num_embeds_ada_norm': self.num_embeds_ada_norm, 'norm_num_groups': 32, 'sample_size': width, 'activation_fn': 'geglu-approximate', } __lowerCamelCase = TransformeraDModel(**lowerCamelCase__ ) return model def lowercase_ ( self ) -> List[str]: '''simple docstring''' __lowerCamelCase = 'cpu' __lowerCamelCase = self.dummy_vqvae __lowerCamelCase = self.dummy_text_encoder __lowerCamelCase = self.dummy_tokenizer __lowerCamelCase = self.dummy_transformer __lowerCamelCase = VQDiffusionScheduler(self.num_embed ) __lowerCamelCase = LearnedClassifierFreeSamplingEmbeddings(learnable=lowerCamelCase__ ) __lowerCamelCase = VQDiffusionPipeline( vqvae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , transformer=lowerCamelCase__ , scheduler=lowerCamelCase__ , learned_classifier_free_sampling_embeddings=lowerCamelCase__ , ) __lowerCamelCase = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = 'teddy bear playing in the pool' __lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __lowerCamelCase = pipe([prompt] , generator=lowerCamelCase__ , num_inference_steps=2 , output_type='np' ) __lowerCamelCase = output.images __lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __lowerCamelCase = pipe( [prompt] , generator=lowerCamelCase__ , output_type='np' , return_dict=lowerCamelCase__ , num_inference_steps=2 )[0] __lowerCamelCase = image[0, -3:, -3:, -1] __lowerCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) __lowerCamelCase = np.array([0.65_51, 0.61_68, 0.50_08, 0.56_76, 0.56_59, 0.42_95, 0.60_73, 0.55_99, 0.49_92] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = 'cpu' __lowerCamelCase = self.dummy_vqvae __lowerCamelCase = self.dummy_text_encoder __lowerCamelCase = self.dummy_tokenizer __lowerCamelCase = self.dummy_transformer __lowerCamelCase = VQDiffusionScheduler(self.num_embed ) __lowerCamelCase = LearnedClassifierFreeSamplingEmbeddings( learnable=lowerCamelCase__ , hidden_size=self.text_embedder_hidden_size , length=tokenizer.model_max_length ) __lowerCamelCase = VQDiffusionPipeline( vqvae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , transformer=lowerCamelCase__ , scheduler=lowerCamelCase__ , learned_classifier_free_sampling_embeddings=lowerCamelCase__ , ) __lowerCamelCase = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = 'teddy bear playing in the pool' __lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __lowerCamelCase = pipe([prompt] , generator=lowerCamelCase__ , num_inference_steps=2 , output_type='np' ) __lowerCamelCase = output.images __lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __lowerCamelCase = pipe( [prompt] , generator=lowerCamelCase__ , output_type='np' , return_dict=lowerCamelCase__ , num_inference_steps=2 )[0] __lowerCamelCase = image[0, -3:, -3:, -1] __lowerCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) __lowerCamelCase = np.array([0.66_93, 0.60_75, 0.49_59, 0.57_01, 0.55_83, 0.43_33, 0.61_71, 0.56_84, 0.49_88] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2.0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/vq_diffusion/teddy_bear_pool_classifier_free_sampling.npy' ) __lowerCamelCase = VQDiffusionPipeline.from_pretrained('microsoft/vq-diffusion-ithq' ) __lowerCamelCase = pipeline.to(lowerCamelCase__ ) pipeline.set_progress_bar_config(disable=lowerCamelCase__ ) # requires GPU generator for gumbel softmax # don't use GPU generator in tests though __lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __lowerCamelCase = pipeline( 'teddy bear playing in the pool' , num_images_per_prompt=1 , generator=lowerCamelCase__ , output_type='np' , ) __lowerCamelCase = output.images[0] assert image.shape == (256, 256, 3) assert np.abs(expected_image - image ).max() < 2.0
348
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a: Any = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a: List[Any] = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a: Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
198
"""simple docstring""" # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys SCREAMING_SNAKE_CASE : Union[str, Any] = subprocess.check_output("""git merge-base main HEAD""".split()).decode("""utf-8""") SCREAMING_SNAKE_CASE : Any = subprocess.check_output(F'git diff --name-only {fork_point_sha}'.split()).decode("""utf-8""").split() SCREAMING_SNAKE_CASE : Union[str, Any] = """|""".join(sys.argv[1:]) SCREAMING_SNAKE_CASE : int = re.compile(rF'^({joined_dirs}).*?\.py$') SCREAMING_SNAKE_CASE : str = [x for x in modified_files if regex.match(x)] print(""" """.join(relevant_modified_files), end="""""")
102
0
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' return [ int(1000 * (box[0] / width) ), int(1000 * (box[1] / height) ), int(1000 * (box[2] / width) ), int(1000 * (box[3] / height) ), ] def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None ) -> str: '''simple docstring''' UpperCamelCase = tesseract_config if tesseract_config is not None else """""" # apply OCR UpperCamelCase = to_pil_image(UpperCamelCase_ ) UpperCamelCase , UpperCamelCase = pil_image.size UpperCamelCase = pytesseract.image_to_data(UpperCamelCase_ , lang=UpperCamelCase_ , output_type="""dict""" , config=UpperCamelCase_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = data["""text"""], data["""left"""], data["""top"""], data["""width"""], data["""height"""] # filter empty words and corresponding coordinates UpperCamelCase = [idx for idx, word in enumerate(UpperCamelCase_ ) if not word.strip()] UpperCamelCase = [word for idx, word in enumerate(UpperCamelCase_ ) if idx not in irrelevant_indices] UpperCamelCase = [coord for idx, coord in enumerate(UpperCamelCase_ ) if idx not in irrelevant_indices] UpperCamelCase = [coord for idx, coord in enumerate(UpperCamelCase_ ) if idx not in irrelevant_indices] UpperCamelCase = [coord for idx, coord in enumerate(UpperCamelCase_ ) if idx not in irrelevant_indices] UpperCamelCase = [coord for idx, coord in enumerate(UpperCamelCase_ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format UpperCamelCase = [] for x, y, w, h in zip(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase = [x, y, x + w, y + h] actual_boxes.append(UpperCamelCase_ ) # finally, normalize the bounding boxes UpperCamelCase = [] for box in actual_boxes: normalized_boxes.append(normalize_box(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) ) assert len(UpperCamelCase_ ) == len(UpperCamelCase_ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = ["""pixel_values"""] def __init__( self : Any , lowerCamelCase_ : bool = True , lowerCamelCase_ : Dict[str, int] = None , lowerCamelCase_ : PILImageResampling = PILImageResampling.BILINEAR , lowerCamelCase_ : bool = True , lowerCamelCase_ : Optional[str] = None , lowerCamelCase_ : Optional[str] = "" , **lowerCamelCase_ : int , ): """simple docstring""" super().__init__(**lowerCamelCase_ ) UpperCamelCase = size if size is not None else {"""height""": 224, """width""": 224} UpperCamelCase = get_size_dict(lowerCamelCase_ ) UpperCamelCase = do_resize UpperCamelCase = size UpperCamelCase = resample UpperCamelCase = apply_ocr UpperCamelCase = ocr_lang UpperCamelCase = tesseract_config def lowerCamelCase_ ( self : Dict , lowerCamelCase_ : np.ndarray , lowerCamelCase_ : Dict[str, int] , lowerCamelCase_ : PILImageResampling = PILImageResampling.BILINEAR , lowerCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ : Any , ): """simple docstring""" UpperCamelCase = get_size_dict(lowerCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) UpperCamelCase = (size["""height"""], size["""width"""]) return resize(lowerCamelCase_ , size=lowerCamelCase_ , resample=lowerCamelCase_ , data_format=lowerCamelCase_ , **lowerCamelCase_ ) def lowerCamelCase_ ( self : int , lowerCamelCase_ : ImageInput , lowerCamelCase_ : bool = None , lowerCamelCase_ : Dict[str, int] = None , lowerCamelCase_ : PILImageResampling = None , lowerCamelCase_ : bool = None , lowerCamelCase_ : Optional[str] = None , lowerCamelCase_ : Optional[str] = None , lowerCamelCase_ : Optional[Union[str, TensorType]] = None , lowerCamelCase_ : ChannelDimension = ChannelDimension.FIRST , **lowerCamelCase_ : int , ): """simple docstring""" UpperCamelCase = do_resize if do_resize is not None else self.do_resize UpperCamelCase = size if size is not None else self.size UpperCamelCase = get_size_dict(lowerCamelCase_ ) UpperCamelCase = resample if resample is not None else self.resample UpperCamelCase = apply_ocr if apply_ocr is not None else self.apply_ocr UpperCamelCase = ocr_lang if ocr_lang is not None else self.ocr_lang UpperCamelCase = tesseract_config if tesseract_config is not None else self.tesseract_config UpperCamelCase = make_list_of_images(lowerCamelCase_ ) if not valid_images(lowerCamelCase_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) # All transformations expect numpy arrays. UpperCamelCase = [to_numpy_array(lowerCamelCase_ ) for image in images] if apply_ocr: requires_backends(self , """pytesseract""" ) UpperCamelCase = [] UpperCamelCase = [] for image in images: UpperCamelCase , UpperCamelCase = apply_tesseract(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) words_batch.append(lowerCamelCase_ ) boxes_batch.append(lowerCamelCase_ ) if do_resize: UpperCamelCase = [self.resize(image=lowerCamelCase_ , size=lowerCamelCase_ , resample=lowerCamelCase_ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) UpperCamelCase = [flip_channel_order(lowerCamelCase_ ) for image in images] UpperCamelCase = [to_channel_dimension_format(lowerCamelCase_ , lowerCamelCase_ ) for image in images] UpperCamelCase = BatchFeature(data={"""pixel_values""": images} , tensor_type=lowerCamelCase_ ) if apply_ocr: UpperCamelCase = words_batch UpperCamelCase = boxes_batch return data
165
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = TypeVar("""DatasetType""", Dataset, IterableDataset) def lowercase( UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = "first_exhausted" , ) -> DatasetType: '''simple docstring''' from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError("""Unable to interleave an empty list of datasets.""" ) for i, dataset in enumerate(UpperCamelCase_ ): if not isinstance(UpperCamelCase_ , (Dataset, IterableDataset) ): if isinstance(UpperCamelCase_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ """is an empty dataset dictionary.""" ) raise ValueError( f"""Dataset at position {i} has at least one split: {list(UpperCamelCase_ )}\n""" f"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(UpperCamelCase_ ) )}']""" ) raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(UpperCamelCase_ ).__name__}.""" ) if i == 0: UpperCamelCase , UpperCamelCase = ( (Dataset, IterableDataset) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else (IterableDataset, Dataset) ) elif not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError( f"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(f"""{stopping_strategy} is not supported. Please enter a valid stopping_strategy.""" ) if dataset_type is Dataset: return _interleave_map_style_datasets( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , stopping_strategy=UpperCamelCase_ ) else: return _interleave_iterable_datasets( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , stopping_strategy=UpperCamelCase_ ) def lowercase( UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = 0 , ) -> DatasetType: '''simple docstring''' if not dsets: raise ValueError("""Unable to concatenate an empty list of datasets.""" ) for i, dataset in enumerate(UpperCamelCase_ ): if not isinstance(UpperCamelCase_ , (Dataset, IterableDataset) ): if isinstance(UpperCamelCase_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} """ """is an empty dataset dictionary.""" ) raise ValueError( f"""Dataset at position {i} has at least one split: {list(UpperCamelCase_ )}\n""" f"""Please pick one to interleave with the other datasets, for example: dataset['{next(iter(UpperCamelCase_ ) )}']""" ) raise ValueError( f"""Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(UpperCamelCase_ ).__name__}.""" ) if i == 0: UpperCamelCase , UpperCamelCase = ( (Dataset, IterableDataset) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else (IterableDataset, Dataset) ) elif not isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError( f"""Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.""" ) if dataset_type is Dataset: return _concatenate_map_style_datasets(UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , axis=UpperCamelCase_ ) else: return _concatenate_iterable_datasets(UpperCamelCase_ , info=UpperCamelCase_ , split=UpperCamelCase_ , axis=UpperCamelCase_ )
165
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase__ = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = ["ChineseCLIPFeatureExtractor"] UpperCAmelCase__ = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
5
class _lowercase : """simple docstring""" def __init__( self : Any , __lowerCamelCase : int ): '''simple docstring''' lowerCamelCase__ : List[str] = n lowerCamelCase__ : Union[str, Any] = [None] * self.n lowerCamelCase__ : List[str] = 0 # index of the first element lowerCamelCase__ : Any = 0 lowerCamelCase__ : Any = 0 def __len__( self : Tuple ): '''simple docstring''' return self.size def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' return self.size == 0 def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' return False if self.is_empty() else self.array[self.front] def lowerCAmelCase ( self : str , __lowerCamelCase : List[str] ): '''simple docstring''' if self.size >= self.n: raise Exception("QUEUE IS FULL" ) lowerCamelCase__ : Optional[Any] = data lowerCamelCase__ : Tuple = (self.rear + 1) % self.n self.size += 1 return self def lowerCAmelCase ( self : Tuple ): '''simple docstring''' if self.size == 0: raise Exception("UNDERFLOW" ) lowerCamelCase__ : Any = self.array[self.front] lowerCamelCase__ : List[Any] = None lowerCamelCase__ : str = (self.front + 1) % self.n self.size -= 1 return temp
184
0
'''simple docstring''' import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowerCAmelCase :Any = random.Random() def lowerCamelCase ( lowerCAmelCase : Tuple , lowerCAmelCase : int=1.0 , lowerCAmelCase : int=None , lowerCAmelCase : List[Any]=None ): """simple docstring""" if rng is None: __magic_name__ : List[str] = global_rng __magic_name__ : List[str] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , _A : int , _A : str=7 , _A : Optional[Any]=400 , _A : List[Any]=2000 , _A : Optional[Any]=1 , _A : Optional[Any]=0.0 , _A : int=16000 , _A : int=True , _A : List[str]=True , ) -> int: __magic_name__ : Optional[int] = parent __magic_name__ : Dict = batch_size __magic_name__ : str = min_seq_length __magic_name__ : Tuple = max_seq_length __magic_name__ : List[Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __magic_name__ : Optional[int] = feature_size __magic_name__ : List[Any] = padding_value __magic_name__ : Tuple = sampling_rate __magic_name__ : List[str] = return_attention_mask __magic_name__ : Union[str, Any] = do_normalize def __lowerCAmelCase ( self : int ) -> int: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def __lowerCAmelCase ( self : List[Any] , _A : List[str]=False , _A : Union[str, Any]=False ) -> str: def _flatten(_A : int ): return list(itertools.chain(*_A ) ) if equal_length: __magic_name__ : Dict = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __magic_name__ : Tuple = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __magic_name__ : Optional[int] = [np.asarray(_A ) for x in speech_inputs] return speech_inputs class _lowerCamelCase ( lowercase__ , unittest.TestCase ): '''simple docstring''' A_ : List[str] = WavaVecaFeatureExtractor def __lowerCAmelCase ( self : str ) -> Optional[int]: __magic_name__ : str = WavaVecaFeatureExtractionTester(self ) def __lowerCAmelCase ( self : Optional[Any] , _A : List[Any] ) -> Any: self.assertTrue(np.all(np.mean(_A , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(_A , axis=0 ) - 1 ) < 1E-3 ) ) def __lowerCAmelCase ( self : List[Any] ) -> Optional[Any]: # Tests that all call wrap to encode_plus and batch_encode_plus __magic_name__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __magic_name__ : Union[str, Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __magic_name__ : Union[str, Any] = [np.asarray(_A ) for speech_input in speech_inputs] # Test not batched input __magic_name__ : List[Any] = feat_extract(speech_inputs[0] , return_tensors='np' ).input_values __magic_name__ : Optional[Any] = feat_extract(np_speech_inputs[0] , return_tensors='np' ).input_values self.assertTrue(np.allclose(_A , _A , atol=1E-3 ) ) # Test batched __magic_name__ : Tuple = feat_extract(_A , return_tensors='np' ).input_values __magic_name__ : str = feat_extract(_A , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(_A , _A ): self.assertTrue(np.allclose(_A , _A , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __magic_name__ : Optional[int] = [floats_list((1, x) )[0] for x in (800, 800, 800)] __magic_name__ : Optional[int] = np.asarray(_A ) __magic_name__ : List[Any] = feat_extract(_A , return_tensors='np' ).input_values __magic_name__ : List[Any] = feat_extract(_A , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(_A , _A ): self.assertTrue(np.allclose(_A , _A , atol=1E-3 ) ) def __lowerCAmelCase ( self : List[Any] ) -> Tuple: __magic_name__ : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __magic_name__ : int = ['longest', 'max_length', 'do_not_pad'] __magic_name__ : Any = [None, 1600, None] for max_length, padding in zip(_A , _A ): __magic_name__ : List[str] = feat_extract(_A , padding=_A , max_length=_A , return_tensors='np' ) __magic_name__ : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __lowerCAmelCase ( self : Optional[int] ) -> List[str]: __magic_name__ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ : List[str] = range(800 , 1400 , 200 ) __magic_name__ : Optional[Any] = [floats_list((1, x) )[0] for x in lengths] __magic_name__ : Tuple = ['longest', 'max_length', 'do_not_pad'] __magic_name__ : Any = [None, 1600, None] for max_length, padding in zip(_A , _A ): __magic_name__ : str = feat_extract(_A , max_length=_A , padding=_A ) __magic_name__ : List[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __lowerCAmelCase ( self : List[Any] ) -> Dict: __magic_name__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ : Any = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __magic_name__ : Optional[Any] = feat_extract( _A , truncation=_A , max_length=1000 , padding='max_length' , return_tensors='np' ) __magic_name__ : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __lowerCAmelCase ( self : Any ) -> Dict: __magic_name__ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ : List[str] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __magic_name__ : Optional[int] = feat_extract( _A , truncation=_A , max_length=1000 , padding='longest' , return_tensors='np' ) __magic_name__ : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) __magic_name__ : Optional[int] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __magic_name__ : str = feat_extract( _A , truncation=_A , max_length=2000 , padding='longest' , return_tensors='np' ) __magic_name__ : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def __lowerCAmelCase ( self : Optional[int] ) -> Any: import torch __magic_name__ : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ : List[str] = np.random.rand(100 ).astype(np.floataa ) __magic_name__ : Union[str, Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __magic_name__ : Optional[int] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __magic_name__ : Dict = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def __lowerCAmelCase ( self : int ) -> int: # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __magic_name__ : Any = WavaVecaConfig.from_pretrained(_A ) __magic_name__ : str = WavaVecaFeatureExtractor.from_pretrained(_A ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer' )
275
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class _lowerCamelCase ( lowercase__ ): '''simple docstring''' A_ : Any = """Salesforce/blip-image-captioning-base""" A_ : Any = ( """This is a tool that generates a description of an image. It takes an input named `image` which should be the """ """image to caption, and returns a text that contains the description in English.""" ) A_ : List[str] = """image_captioner""" A_ : Optional[Any] = AutoModelForVisionaSeq A_ : Any = ["""image"""] A_ : Union[str, Any] = ["""text"""] def __init__( self : Dict , *_A : Dict , **_A : Any ) -> Dict: requires_backends(self , ['vision'] ) super().__init__(*_A , **_A ) def __lowerCAmelCase ( self : Tuple , _A : "Image" ) -> Union[str, Any]: return self.pre_processor(images=_A , return_tensors='pt' ) def __lowerCAmelCase ( self : Optional[Any] , _A : str ) -> List[Any]: return self.model.generate(**_A ) def __lowerCAmelCase ( self : int , _A : Optional[int] ) -> Union[str, Any]: return self.pre_processor.batch_decode(_A , skip_special_tokens=_A )[0].strip()
275
1
from typing import TYPE_CHECKING from ...utils import _LazyModule lowerCamelCase__ = {"""tokenization_bertweet""": ["""BertweetTokenizer"""]} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
302
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration lowerCAmelCase_ = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def lowerCamelCase_ ( _UpperCamelCase ) -> List[Any]: """simple docstring""" snake_case_ : List[str] = ['''layers''', '''blocks'''] for k in ignore_keys: state_dict.pop(_UpperCamelCase , _UpperCamelCase ) lowerCAmelCase_ = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" snake_case_ : str = list(s_dict.keys() ) for key in keys: snake_case_ : Optional[int] = key for k, v in WHISPER_MAPPING.items(): if k in key: snake_case_ : List[str] = new_key.replace(_UpperCamelCase , _UpperCamelCase ) print(f'''{key} -> {new_key}''' ) snake_case_ : Tuple = s_dict.pop(_UpperCamelCase ) return s_dict def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" snake_case_ , snake_case_ : Dict = emb.weight.shape snake_case_ : Tuple = nn.Linear(_UpperCamelCase , _UpperCamelCase , bias=_UpperCamelCase ) snake_case_ : Any = emb.weight.data return lin_layer def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> bytes: """simple docstring""" os.makedirs(_UpperCamelCase , exist_ok=_UpperCamelCase ) snake_case_ : List[Any] = os.path.basename(_UpperCamelCase ) snake_case_ : Any = url.split('''/''' )[-2] snake_case_ : str = os.path.join(_UpperCamelCase , _UpperCamelCase ) if os.path.exists(_UpperCamelCase ) and not os.path.isfile(_UpperCamelCase ): raise RuntimeError(f'''{download_target} exists and is not a regular file''' ) if os.path.isfile(_UpperCamelCase ): snake_case_ : Union[str, Any] = open(_UpperCamelCase , '''rb''' ).read() if hashlib.shaaaa(_UpperCamelCase ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(f'''{download_target} exists, but the SHA256 checksum does not match; re-downloading the file''' ) with urllib.request.urlopen(_UpperCamelCase ) as source, open(_UpperCamelCase , '''wb''' ) as output: with tqdm( total=int(source.info().get('''Content-Length''' ) ) , ncols=80 , unit='''iB''' , unit_scale=_UpperCamelCase , unit_divisor=1_024 ) as loop: while True: snake_case_ : Dict = source.read(8_192 ) if not buffer: break output.write(_UpperCamelCase ) loop.update(len(_UpperCamelCase ) ) snake_case_ : Any = open(_UpperCamelCase , '''rb''' ).read() if hashlib.shaaaa(_UpperCamelCase ).hexdigest() != expected_shaaaa: raise RuntimeError( '''Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.''' ) return model_bytes def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> int: """simple docstring""" if ".pt" not in checkpoint_path: snake_case_ : str = _download(_MODELS[checkpoint_path] ) else: snake_case_ : Union[str, Any] = torch.load(_UpperCamelCase , map_location='''cpu''' ) snake_case_ : int = original_checkpoint['''dims'''] snake_case_ : List[str] = original_checkpoint['''model_state_dict'''] snake_case_ : str = state_dict['''decoder.token_embedding.weight'''] remove_ignore_keys_(_UpperCamelCase ) rename_keys(_UpperCamelCase ) snake_case_ : Optional[int] = True snake_case_ : int = state_dict['''decoder.layers.0.fc1.weight'''].shape[0] snake_case_ : List[str] = WhisperConfig( vocab_size=dimensions['''n_vocab'''] , encoder_ffn_dim=_UpperCamelCase , decoder_ffn_dim=_UpperCamelCase , num_mel_bins=dimensions['''n_mels'''] , d_model=dimensions['''n_audio_state'''] , max_target_positions=dimensions['''n_text_ctx'''] , encoder_layers=dimensions['''n_audio_layer'''] , encoder_attention_heads=dimensions['''n_audio_head'''] , decoder_layers=dimensions['''n_text_layer'''] , decoder_attention_heads=dimensions['''n_text_state'''] , max_source_positions=dimensions['''n_audio_ctx'''] , ) snake_case_ : Union[str, Any] = WhisperForConditionalGeneration(_UpperCamelCase ) snake_case_ , snake_case_ : List[Any] = model.model.load_state_dict(_UpperCamelCase , strict=_UpperCamelCase ) if len(_UpperCamelCase ) > 0 and not set(_UpperCamelCase ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f''' but all the following weights are missing {missing}''' ) if tie_embeds: snake_case_ : List[str] = make_linear_from_emb(model.model.decoder.embed_tokens ) else: snake_case_ : Any = proj_out_weights model.save_pretrained(_UpperCamelCase ) if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') lowerCAmelCase_ = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
279
0
from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError('''To use the rich extension, install rich with `pip install rich`''')
365
from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) def lowerCAmelCase_ ( *__lowerCAmelCase , **__lowerCAmelCase )-> List[str]: '''simple docstring''' requires_backends(__lowerCAmelCase , ['''torch'''] ) def lowerCAmelCase_ ( *__lowerCAmelCase , **__lowerCAmelCase )-> Tuple: '''simple docstring''' requires_backends(__lowerCAmelCase , ['''torch'''] ) def lowerCAmelCase_ ( *__lowerCAmelCase , **__lowerCAmelCase )-> List[str]: '''simple docstring''' requires_backends(__lowerCAmelCase , ['''torch'''] ) def lowerCAmelCase_ ( *__lowerCAmelCase , **__lowerCAmelCase )-> Optional[int]: '''simple docstring''' requires_backends(__lowerCAmelCase , ['''torch'''] ) def lowerCAmelCase_ ( *__lowerCAmelCase , **__lowerCAmelCase )-> Union[str, Any]: '''simple docstring''' requires_backends(__lowerCAmelCase , ['''torch'''] ) def lowerCAmelCase_ ( *__lowerCAmelCase , **__lowerCAmelCase )-> Optional[int]: '''simple docstring''' requires_backends(__lowerCAmelCase , ['''torch'''] ) def lowerCAmelCase_ ( *__lowerCAmelCase , **__lowerCAmelCase )-> List[Any]: '''simple docstring''' requires_backends(__lowerCAmelCase , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Optional[int] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Union[str, Any] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> List[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : int = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Any = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Dict: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Dict = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> str: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Any: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : str = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> int: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : Tuple = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> Optional[int]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(cls , ['''torch'''] ) class __snake_case ( metaclass=lowerCamelCase__ ): __lowerCamelCase : List[str] = ["""torch"""] def __init__( self , *snake_case__ , **snake_case__ ) -> List[str]: '''simple docstring''' requires_backends(self , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Tuple: '''simple docstring''' requires_backends(cls , ['''torch'''] ) @classmethod def UpperCAmelCase__ ( cls , *snake_case__ , **snake_case__ ) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['''torch'''] )
78
0
'''simple docstring''' import math def a_ ( __snake_case : list , __snake_case : int = 0 , __snake_case : int = 0 ) -> list: """simple docstring""" lowerCamelCase_ =end or len(__snake_case ) for i in range(__snake_case , __snake_case ): lowerCamelCase_ =i lowerCamelCase_ =array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: lowerCamelCase_ =array[temp_index - 1] temp_index -= 1 lowerCamelCase_ =temp_index_value return array def a_ ( __snake_case : list , __snake_case : int , __snake_case : int ) -> None: # Max Heap """simple docstring""" lowerCamelCase_ =index lowerCamelCase_ =2 * index + 1 # Left Node lowerCamelCase_ =2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: lowerCamelCase_ =left_index if right_index < heap_size and array[largest] < array[right_index]: lowerCamelCase_ =right_index if largest != index: lowerCamelCase_, lowerCamelCase_ =array[largest], array[index] heapify(__snake_case , __snake_case , __snake_case ) def a_ ( __snake_case : list ) -> list: """simple docstring""" lowerCamelCase_ =len(__snake_case ) for i in range(n // 2 , -1 , -1 ): heapify(__snake_case , __snake_case , __snake_case ) for i in range(n - 1 , 0 , -1 ): lowerCamelCase_, lowerCamelCase_ =array[0], array[i] heapify(__snake_case , 0 , __snake_case ) return array def a_ ( __snake_case : list , __snake_case : int , __snake_case : int , __snake_case : int ) -> int: """simple docstring""" if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def a_ ( __snake_case : list , __snake_case : int , __snake_case : int , __snake_case : int ) -> int: """simple docstring""" lowerCamelCase_ =low lowerCamelCase_ =high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i lowerCamelCase_, lowerCamelCase_ =array[j], array[i] i += 1 def a_ ( __snake_case : list ) -> list: """simple docstring""" if len(__snake_case ) == 0: return array lowerCamelCase_ =2 * math.ceil(math.loga(len(__snake_case ) ) ) lowerCamelCase_ =16 return intro_sort(__snake_case , 0 , len(__snake_case ) , __snake_case , __snake_case ) def a_ ( __snake_case : list , __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : int ) -> list: """simple docstring""" while end - start > size_threshold: if max_depth == 0: return heap_sort(__snake_case ) max_depth -= 1 lowerCamelCase_ =median_of_a(__snake_case , __snake_case , start + ((end - start) // 2) + 1 , end - 1 ) lowerCamelCase_ =partition(__snake_case , __snake_case , __snake_case , __snake_case ) intro_sort(__snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) lowerCamelCase_ =p return insertion_sort(__snake_case , __snake_case , __snake_case ) if __name__ == "__main__": import doctest doctest.testmod() a_ : Optional[Any] = input("""Enter numbers separated by a comma : """).strip() a_ : List[Any] = [float(item) for item in user_input.split(""",""")] print(sort(unsorted))
75
def __lowerCAmelCase ( a__ , a__ , a__ ) -> list: __a = len(a__ ) __a = [[0] * n for i in range(a__ )] for i in range(a__ ): __a = y_points[i] for i in range(2 , a__ ): for j in range(a__ , a__ ): __a = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
6
0
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class lowerCamelCase : '''simple docstring''' def __init__( self , _UpperCamelCase , _UpperCamelCase=1_3 , _UpperCamelCase=7 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=9_9 , _UpperCamelCase=3_2 , _UpperCamelCase=2 , _UpperCamelCase=4 , _UpperCamelCase=3_7 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=5_1_2 , _UpperCamelCase=1_6 , _UpperCamelCase=2 , _UpperCamelCase=0.02 , _UpperCamelCase=3 , _UpperCamelCase=4 , _UpperCamelCase=None , ) -> List[Any]: UpperCAmelCase_ : List[Any] = parent UpperCAmelCase_ : List[Any] = 1_3 UpperCAmelCase_ : str = 7 UpperCAmelCase_ : str = True UpperCAmelCase_ : Union[str, Any] = True UpperCAmelCase_ : Optional[int] = True UpperCAmelCase_ : Dict = True UpperCAmelCase_ : Optional[Any] = 9_9 UpperCAmelCase_ : Optional[Any] = 3_8_4 UpperCAmelCase_ : List[str] = 2 UpperCAmelCase_ : str = 4 UpperCAmelCase_ : Optional[Any] = 3_7 UpperCAmelCase_ : Dict = 'gelu' UpperCAmelCase_ : List[str] = 0.1 UpperCAmelCase_ : List[str] = 0.1 UpperCAmelCase_ : Any = 5_1_2 UpperCAmelCase_ : str = 1_6 UpperCAmelCase_ : Optional[int] = 2 UpperCAmelCase_ : str = 0.02 UpperCAmelCase_ : Tuple = 3 UpperCAmelCase_ : Dict = 4 UpperCAmelCase_ : Optional[int] = 1_2_8 UpperCAmelCase_ : List[str] = 2 UpperCAmelCase_ : Optional[int] = 9 UpperCAmelCase_ : Union[str, Any] = 1 UpperCAmelCase_ : Union[str, Any] = None def __UpperCAmelCase ( self ) -> Tuple: UpperCAmelCase_ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_ : List[str] = None if self.use_input_mask: UpperCAmelCase_ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ : Any = None if self.use_token_type_ids: UpperCAmelCase_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase_ : List[str] = None UpperCAmelCase_ : Any = None UpperCAmelCase_ : Any = None if self.use_labels: UpperCAmelCase_ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase_ : List[str] = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase_ : Optional[int] = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=_UpperCamelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Tuple: UpperCAmelCase_ : Optional[Any] = TFConvBertModel(config=_UpperCamelCase ) UpperCAmelCase_ : List[str] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} UpperCAmelCase_ : Optional[Any] = [input_ids, input_mask] UpperCAmelCase_ : Dict = model(_UpperCamelCase ) UpperCAmelCase_ : List[str] = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Union[str, Any]: UpperCAmelCase_ : Dict = TFConvBertForMaskedLM(config=_UpperCamelCase ) UpperCAmelCase_ : Dict = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase_ : Dict = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> List[str]: UpperCAmelCase_ : str = self.num_labels UpperCAmelCase_ : Dict = TFConvBertForSequenceClassification(config=_UpperCamelCase ) UpperCAmelCase_ : int = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase_ : List[Any] = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: UpperCAmelCase_ : Any = self.num_choices UpperCAmelCase_ : Union[str, Any] = TFConvBertForMultipleChoice(config=_UpperCamelCase ) UpperCAmelCase_ : Union[str, Any] = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase_ : Dict = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase_ : List[str] = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase_ : Union[str, Any] = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } UpperCAmelCase_ : List[str] = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Tuple: UpperCAmelCase_ : Union[str, Any] = self.num_labels UpperCAmelCase_ : str = TFConvBertForTokenClassification(config=_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase_ : Any = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Union[str, Any]: UpperCAmelCase_ : str = TFConvBertForQuestionAnswering(config=_UpperCamelCase ) UpperCAmelCase_ : str = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } UpperCAmelCase_ : Tuple = model(_UpperCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCAmelCase ( self ) -> Any: UpperCAmelCase_ : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) : List[Any] = config_and_inputs UpperCAmelCase_ : Optional[int] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class lowerCamelCase (_snake_case , _snake_case , unittest.TestCase ): '''simple docstring''' _snake_case : str = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) _snake_case : Tuple = ( { '''feature-extraction''': TFConvBertModel, '''fill-mask''': TFConvBertForMaskedLM, '''question-answering''': TFConvBertForQuestionAnswering, '''text-classification''': TFConvBertForSequenceClassification, '''token-classification''': TFConvBertForTokenClassification, '''zero-shot''': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) _snake_case : List[Any] = False _snake_case : Union[str, Any] = False _snake_case : str = False def __UpperCAmelCase ( self ) -> List[Any]: UpperCAmelCase_ : Union[str, Any] = TFConvBertModelTester(self ) UpperCAmelCase_ : Dict = ConfigTester(self , config_class=_UpperCamelCase , hidden_size=3_7 ) def __UpperCAmelCase ( self ) -> List[Any]: self.config_tester.run_common_tests() def __UpperCAmelCase ( self ) -> Any: UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCamelCase ) def __UpperCAmelCase ( self ) -> int: UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_UpperCamelCase ) def __UpperCAmelCase ( self ) -> str: UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Optional[Any]: UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_UpperCamelCase ) def __UpperCAmelCase ( self ) -> int: UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Optional[Any]: UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_UpperCamelCase ) @slow def __UpperCAmelCase ( self ) -> Tuple: UpperCAmelCase_ , UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Optional[int] = True UpperCAmelCase_ : Tuple = True if hasattr(_UpperCamelCase , 'use_cache' ): UpperCAmelCase_ : List[str] = True UpperCAmelCase_ : List[Any] = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) UpperCAmelCase_ : Optional[int] = getattr(self.model_tester , 'key_length' , _UpperCamelCase ) for model_class in self.all_model_classes: UpperCAmelCase_ : Union[str, Any] = self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) UpperCAmelCase_ : Dict = model_class(_UpperCamelCase ) UpperCAmelCase_ : List[Any] = len(model(_UpperCamelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_UpperCamelCase , saved_model=_UpperCamelCase ) UpperCAmelCase_ : List[str] = os.path.join(_UpperCamelCase , 'saved_model' , '1' ) UpperCAmelCase_ : Tuple = tf.keras.models.load_model(_UpperCamelCase ) UpperCAmelCase_ : List[str] = model(_UpperCamelCase ) if self.is_encoder_decoder: UpperCAmelCase_ : str = outputs['encoder_hidden_states'] UpperCAmelCase_ : List[Any] = outputs['encoder_attentions'] else: UpperCAmelCase_ : str = outputs['hidden_states'] UpperCAmelCase_ : List[Any] = outputs['attentions'] self.assertEqual(len(_UpperCamelCase ) , _UpperCamelCase ) UpperCAmelCase_ : str = getattr( self.model_tester , 'expected_num_hidden_layers' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_UpperCamelCase ) , _UpperCamelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __UpperCAmelCase ( self ) -> Union[str, Any]: UpperCAmelCase_ : Tuple = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) self.assertIsNotNone(_UpperCamelCase ) def __UpperCAmelCase ( self ) -> Any: UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Any = True UpperCAmelCase_ : List[Any] = getattr(self.model_tester , 'decoder_seq_length' , self.model_tester.seq_length ) UpperCAmelCase_ : Optional[int] = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) UpperCAmelCase_ : Dict = getattr(self.model_tester , 'key_length' , _UpperCamelCase ) UpperCAmelCase_ : int = getattr(self.model_tester , 'key_length' , _UpperCamelCase ) def check_decoder_attentions_output(_UpperCamelCase ): UpperCAmelCase_ : int = len(_UpperCamelCase ) self.assertEqual(out_len % 2 , 0 ) UpperCAmelCase_ : List[Any] = outputs.decoder_attentions self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(_UpperCamelCase ): UpperCAmelCase_ : Union[str, Any] = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(_UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCAmelCase_ : Tuple = True UpperCAmelCase_ : List[Any] = False UpperCAmelCase_ : Optional[int] = model_class(_UpperCamelCase ) UpperCAmelCase_ : Union[str, Any] = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) UpperCAmelCase_ : Union[str, Any] = len(_UpperCamelCase ) self.assertEqual(config.output_hidden_states , _UpperCamelCase ) check_encoder_attentions_output(_UpperCamelCase ) if self.is_encoder_decoder: UpperCAmelCase_ : Any = model_class(_UpperCamelCase ) UpperCAmelCase_ : Tuple = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCamelCase ) check_decoder_attentions_output(_UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCAmelCase_ : str = True UpperCAmelCase_ : Tuple = model_class(_UpperCamelCase ) UpperCAmelCase_ : Union[str, Any] = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCamelCase ) check_encoder_attentions_output(_UpperCamelCase ) # Check attention is always last and order is fine UpperCAmelCase_ : List[Any] = True UpperCAmelCase_ : Tuple = True UpperCAmelCase_ : Optional[int] = model_class(_UpperCamelCase ) UpperCAmelCase_ : Dict = model(self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states , _UpperCamelCase ) check_encoder_attentions_output(_UpperCamelCase ) @require_tf class lowerCamelCase (unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self ) -> Optional[Any]: UpperCAmelCase_ : Optional[int] = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) UpperCAmelCase_ : Any = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase_ : List[str] = model(_UpperCamelCase )[0] UpperCAmelCase_ : Dict = [1, 6, 7_6_8] self.assertEqual(output.shape , _UpperCamelCase ) UpperCAmelCase_ : Optional[int] = tf.constant( [ [ [-0.03_47_54_93, -0.4_68_60_34, -0.30_63_88_32], [0.22_63_72_48, -0.26_98_86_46, -0.7_42_34_24], [0.10_32_48_68, -0.45_01_35_08, -0.58_28_07_84], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , _UpperCamelCase , atol=1E-4 )
145
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'microsoft/layoutlmv3-base': 'https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json', } class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Dict = '''layoutlmv3''' def __init__( self , _UpperCamelCase=5_0_2_6_5 , _UpperCamelCase=7_6_8 , _UpperCamelCase=1_2 , _UpperCamelCase=1_2 , _UpperCamelCase=3_0_7_2 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=5_1_2 , _UpperCamelCase=2 , _UpperCamelCase=0.02 , _UpperCamelCase=1E-5 , _UpperCamelCase=1 , _UpperCamelCase=0 , _UpperCamelCase=2 , _UpperCamelCase=1_0_2_4 , _UpperCamelCase=1_2_8 , _UpperCamelCase=1_2_8 , _UpperCamelCase=True , _UpperCamelCase=3_2 , _UpperCamelCase=1_2_8 , _UpperCamelCase=6_4 , _UpperCamelCase=2_5_6 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=2_2_4 , _UpperCamelCase=3 , _UpperCamelCase=1_6 , _UpperCamelCase=None , **_UpperCamelCase , ) -> Optional[Any]: super().__init__( vocab_size=_UpperCamelCase , hidden_size=_UpperCamelCase , num_hidden_layers=_UpperCamelCase , num_attention_heads=_UpperCamelCase , intermediate_size=_UpperCamelCase , hidden_act=_UpperCamelCase , hidden_dropout_prob=_UpperCamelCase , attention_probs_dropout_prob=_UpperCamelCase , max_position_embeddings=_UpperCamelCase , type_vocab_size=_UpperCamelCase , initializer_range=_UpperCamelCase , layer_norm_eps=_UpperCamelCase , pad_token_id=_UpperCamelCase , bos_token_id=_UpperCamelCase , eos_token_id=_UpperCamelCase , **_UpperCamelCase , ) UpperCAmelCase_ : str = max_ad_position_embeddings UpperCAmelCase_ : Union[str, Any] = coordinate_size UpperCAmelCase_ : Union[str, Any] = shape_size UpperCAmelCase_ : str = has_relative_attention_bias UpperCAmelCase_ : Tuple = rel_pos_bins UpperCAmelCase_ : Dict = max_rel_pos UpperCAmelCase_ : Any = has_spatial_attention_bias UpperCAmelCase_ : Optional[Any] = rel_ad_pos_bins UpperCAmelCase_ : List[str] = max_rel_ad_pos UpperCAmelCase_ : List[str] = text_embed UpperCAmelCase_ : Dict = visual_embed UpperCAmelCase_ : Optional[int] = input_size UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : List[Any] = patch_size UpperCAmelCase_ : Union[str, Any] = classifier_dropout class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Tuple = version.parse('''1.12''' ) @property def __UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: # The order of inputs is different for question answering and sequence classification if self.task in ["question-answering", "sequence-classification"]: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) else: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels'}), ] ) @property def __UpperCAmelCase ( self ) -> float: return 1E-5 @property def __UpperCAmelCase ( self ) -> int: return 1_2 def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase = -1 , _UpperCamelCase = -1 , _UpperCamelCase = False , _UpperCamelCase = None , _UpperCamelCase = 3 , _UpperCamelCase = 4_0 , _UpperCamelCase = 4_0 , ) -> Mapping[str, Any]: setattr(processor.image_processor , 'apply_ocr' , _UpperCamelCase ) # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase_ : Optional[int] = compute_effective_axis_dimension( _UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase_ : Any = processor.tokenizer.num_special_tokens_to_add(_UpperCamelCase ) UpperCAmelCase_ : Any = compute_effective_axis_dimension( _UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_UpperCamelCase ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase_ : Tuple = [[' '.join([processor.tokenizer.unk_token] ) * seq_length]] * batch_size # Generate dummy bounding boxes UpperCAmelCase_ : List[str] = [[[4_8, 8_4, 7_3, 1_2_8]]] * batch_size # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX # batch_size = compute_effective_axis_dimension(batch_size, fixed_dimension=OnnxConfig.default_fixed_batch) UpperCAmelCase_ : str = self._generate_dummy_images(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) UpperCAmelCase_ : List[str] = dict( processor( _UpperCamelCase , text=_UpperCamelCase , boxes=_UpperCamelCase , return_tensors=_UpperCamelCase , ) ) return inputs
145
1
import qiskit def __UpperCamelCase ( _A , _A ): lowerCAmelCase_ = qiskit.Aer.get_backend('''aer_simulator''' ) # Create a Quantum Circuit acting on the q register lowerCAmelCase_ = qiskit.QuantumCircuit(_A , _A ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator lowerCAmelCase_ = qiskit.execute(_A , _A , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_A ) if __name__ == "__main__": print(f"Total count for various states are: {single_qubit_measure(1, 1)}")
278
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { '''google/vit-base-patch16-224''': '''https://huggingface.co/vit-base-patch16-224/resolve/main/config.json''', # See all ViT models at https://huggingface.co/models?filter=vit } class A ( __UpperCAmelCase ): __snake_case = 'vit' def __init__( self, UpperCamelCase__=768, UpperCamelCase__=12, UpperCamelCase__=12, UpperCamelCase__=3072, UpperCamelCase__="gelu", UpperCamelCase__=0.0, UpperCamelCase__=0.0, UpperCamelCase__=0.02, UpperCamelCase__=1E-12, UpperCamelCase__=224, UpperCamelCase__=16, UpperCamelCase__=3, UpperCamelCase__=True, UpperCamelCase__=16, **UpperCamelCase__, ): """simple docstring""" super().__init__(**UpperCamelCase__ ) lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_act lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = initializer_range lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = image_size lowerCAmelCase_ = patch_size lowerCAmelCase_ = num_channels lowerCAmelCase_ = qkv_bias lowerCAmelCase_ = encoder_stride class A ( __UpperCAmelCase ): __snake_case = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return 1E-4
278
1
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder _lowerCAmelCase : Optional[Any] = '''base_with_context''' def __snake_case ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> int: A_ : Tuple = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) A_ : List[str] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=_lowerCAmelCase ) for lyr_num, lyr in enumerate(model.encoders ): A_ : Dict = weights[f"layers_{lyr_num}"] A_ : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) A_ : Optional[int] = ly_weight["attention"] A_ : Any = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) A_ : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) A_ : Any = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def __snake_case ( _lowerCAmelCase : int , _lowerCAmelCase : List[str] ) -> List[str]: A_ : Tuple = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) A_ : str = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=_lowerCAmelCase ) for lyr_num, lyr in enumerate(model.encoders ): A_ : str = weights[f"layers_{lyr_num}"] A_ : Dict = ly_weight["attention"] A_ : Any = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) A_ : int = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) A_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) A_ : Tuple = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) A_ : Dict = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) A_ : List[str] = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def __snake_case ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] ) -> List[Any]: A_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) A_ : Any = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) A_ : List[Any] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=_lowerCAmelCase ) A_ : Optional[Any] = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): A_ : str = weights[f"layers_{lyr_num}"] A_ : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) A_ : Tuple = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) A_ : List[Any] = ly_weight["self_attention"] A_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) A_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) A_ : Dict = ly_weight["MultiHeadDotProductAttention_0"] A_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) A_ : str = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) A_ : str = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) A_ : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) A_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) A_ : Tuple = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) A_ : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) A_ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) A_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) A_ : int = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) A_ : Optional[int] = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def __snake_case ( _lowerCAmelCase : Optional[int] ) -> Optional[int]: A_ : Tuple = checkpoints.load_tax_checkpoint(args.checkpoint_path ) A_ : Union[str, Any] = jnp.tree_util.tree_map(onp.array , _lowerCAmelCase ) A_ : Tuple = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] A_ : Union[str, Any] = os.path.join(args.checkpoint_path , ".." , "config.gin" ) A_ : List[Any] = inference.parse_training_gin_file(_lowerCAmelCase , _lowerCAmelCase ) A_ : Tuple = inference.InferenceModel(args.checkpoint_path , _lowerCAmelCase ) A_ : int = DDPMScheduler(beta_schedule="squaredcos_cap_v2" , variance_type="fixed_large" ) A_ : Any = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) A_ : Optional[Any] = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length["targets_context"] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) A_ : List[Any] = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length["targets_context"] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) A_ : Union[str, Any] = load_notes_encoder(ta_checkpoint["target"]["token_encoder"] , _lowerCAmelCase ) A_ : Any = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"] , _lowerCAmelCase ) A_ : Union[str, Any] = load_decoder(ta_checkpoint["target"]["decoder"] , _lowerCAmelCase ) A_ : List[Any] = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) A_ : Optional[Any] = SpectrogramDiffusionPipeline( notes_encoder=_lowerCAmelCase , continuous_encoder=_lowerCAmelCase , decoder=_lowerCAmelCase , scheduler=_lowerCAmelCase , melgan=_lowerCAmelCase , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": _lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--output_path''', default=None, type=str, required=True, help='''Path to the converted model.''') parser.add_argument( '''--save''', default=True, type=bool, required=False, help='''Whether to save the converted model or not.''' ) parser.add_argument( '''--checkpoint_path''', default=F'''{MODEL}/checkpoint_500000''', type=str, required=False, help='''Path to the original jax model checkpoint.''', ) _lowerCAmelCase : str = parser.parse_args() main(args)
365
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase : List[Any] = { '''configuration_funnel''': ['''FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FunnelConfig'''], '''convert_funnel_original_tf_checkpoint_to_pytorch''': [], '''tokenization_funnel''': ['''FunnelTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Any = ['''FunnelTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Tuple = [ '''FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FunnelBaseModel''', '''FunnelForMaskedLM''', '''FunnelForMultipleChoice''', '''FunnelForPreTraining''', '''FunnelForQuestionAnswering''', '''FunnelForSequenceClassification''', '''FunnelForTokenClassification''', '''FunnelModel''', '''FunnelPreTrainedModel''', '''load_tf_weights_in_funnel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : int = [ '''TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFFunnelBaseModel''', '''TFFunnelForMaskedLM''', '''TFFunnelForMultipleChoice''', '''TFFunnelForPreTraining''', '''TFFunnelForQuestionAnswering''', '''TFFunnelForSequenceClassification''', '''TFFunnelForTokenClassification''', '''TFFunnelModel''', '''TFFunnelPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_funnel import FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP, FunnelConfig from .tokenization_funnel import FunnelTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_funnel_fast import FunnelTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_funnel import ( FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, FunnelBaseModel, FunnelForMaskedLM, FunnelForMultipleChoice, FunnelForPreTraining, FunnelForQuestionAnswering, FunnelForSequenceClassification, FunnelForTokenClassification, FunnelModel, FunnelPreTrainedModel, load_tf_weights_in_funnel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_funnel import ( TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, TFFunnelPreTrainedModel, ) else: import sys _lowerCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
70
0
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class SCREAMING_SNAKE_CASE__ : def __init__(self : str , a__ : Dict , a__ : Tuple=None , a__ : List[Any]=None , a__ : Dict=None , a__ : Union[str, Any]="resnet50" , a__ : Dict=3 , a__ : str=32 , a__ : int=3 , a__ : Dict=True , a__ : Any=True , ): """simple docstring""" __snake_case = parent __snake_case = out_indices if out_indices is not None else [4] __snake_case = stage_names __snake_case = out_features __snake_case = backbone __snake_case = batch_size __snake_case = image_size __snake_case = num_channels __snake_case = use_pretrained_backbone __snake_case = is_training def a (self : Union[str, Any] ): """simple docstring""" __snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case = self.get_config() return config, pixel_values def a (self : Any ): """simple docstring""" return TimmBackboneConfig( image_size=self.image_size , num_channels=self.num_channels , out_features=self.out_features , out_indices=self.out_indices , stage_names=self.stage_names , use_pretrained_backbone=self.use_pretrained_backbone , backbone=self.backbone , ) def a (self : List[Any] , a__ : int , a__ : int ): """simple docstring""" __snake_case = TimmBackbone(config=a__ ) model.to(a__ ) model.eval() with torch.no_grad(): __snake_case = model(a__ ) self.parent.assertEqual( result.feature_map[-1].shape , (self.batch_size, model.channels[-1], 14, 14) , ) def a (self : str ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() __snake_case , __snake_case = config_and_inputs __snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch @require_timm class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A_ : Union[str, Any] = (TimmBackbone,) if is_torch_available() else () A_ : Optional[Any] = {'feature-extraction': TimmBackbone} if is_torch_available() else {} A_ : List[Any] = False A_ : Dict = False A_ : Any = False A_ : List[Any] = False def a (self : Tuple ): """simple docstring""" __snake_case = TimmBackboneModelTester(self ) __snake_case = ConfigTester(self , config_class=a__ , has_text_modality=a__ ) def a (self : Any ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a (self : int ): """simple docstring""" __snake_case = '''resnet18''' __snake_case = '''microsoft/resnet-18''' __snake_case = AutoBackbone.from_pretrained(a__ , use_timm_backbone=a__ ) __snake_case = AutoBackbone.from_pretrained(a__ ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) , len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices , (-1,) ) self.assertEqual(transformers_model.out_indices , [len(timm_model.stage_names ) - 1] ) __snake_case = AutoBackbone.from_pretrained(a__ , use_timm_backbone=a__ , out_indices=[1, 2, 3] ) __snake_case = AutoBackbone.from_pretrained(a__ , out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices , transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) @unittest.skip('''TimmBackbone doesn\'t support feed forward chunking''' ) def a (self : str ): """simple docstring""" pass @unittest.skip('''TimmBackbone doesn\'t have num_hidden_layers attribute''' ) def a (self : int ): """simple docstring""" pass @unittest.skip('''TimmBackbone initialization is managed on the timm side''' ) def a (self : Union[str, Any] ): """simple docstring""" pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def a (self : Optional[int] ): """simple docstring""" pass @unittest.skip('''TimmBackbone models doesn\'t have inputs_embeds''' ) def a (self : int ): """simple docstring""" pass @unittest.skip('''TimmBackbone model cannot be created without specifying a backbone checkpoint''' ) def a (self : Tuple ): """simple docstring""" pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def a (self : int ): """simple docstring""" pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def a (self : Optional[Any] ): """simple docstring""" pass @unittest.skip('''model weights aren\'t tied in TimmBackbone.''' ) def a (self : Tuple ): """simple docstring""" pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def a (self : Dict ): """simple docstring""" pass @unittest.skip('''Only checkpoints on timm can be loaded into TimmBackbone''' ) def a (self : List[Any] ): """simple docstring""" pass @unittest.skip('''TimmBackbone doesn\'t have hidden size info in its configuration.''' ) def a (self : Optional[Any] ): """simple docstring""" pass @unittest.skip('''TimmBackbone doesn\'t support output_attentions.''' ) def a (self : List[Any] ): """simple docstring""" pass @unittest.skip('''Safetensors is not supported by timm.''' ) def a (self : Tuple ): """simple docstring""" pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def a (self : Tuple ): """simple docstring""" pass def a (self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = model_class(a__ ) __snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case = [*signature.parameters.keys()] __snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , a__ ) def a (self : Dict ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = True __snake_case = self.has_attentions # no need to test all models as different heads yield the same functionality __snake_case = self.all_model_classes[0] __snake_case = model_class(a__ ) model.to(a__ ) __snake_case = self._prepare_for_class(a__ , a__ ) __snake_case = model(**a__ ) __snake_case = outputs[0][-1] # Encoder-/Decoder-only models __snake_case = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: __snake_case = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=a__ ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def a (self : Optional[int] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = model_class(a__ ) model.to(a__ ) model.eval() __snake_case = model(**a__ ) self.assertEqual(len(result.feature_maps ) , len(config.out_indices ) ) self.assertEqual(len(model.channels ) , len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None __snake_case = copy.deepcopy(a__ ) __snake_case = None __snake_case = model_class(a__ ) model.to(a__ ) model.eval() __snake_case = model(**a__ ) self.assertEqual(len(result.feature_maps ) , 1 ) self.assertEqual(len(model.channels ) , 1 ) # Check backbone can be initialized with fresh weights __snake_case = copy.deepcopy(a__ ) __snake_case = False __snake_case = model_class(a__ ) model.to(a__ ) model.eval() __snake_case = model(**a__ )
24
def __lowercase ( __lowerCAmelCase : int , __lowerCAmelCase : float , __lowerCAmelCase : float ): return round(float(moles / volume ) * nfactor ) def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float , __lowerCAmelCase : float ): return round(float((moles * 0.0_821 * temperature) / (volume) ) ) def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float , __lowerCAmelCase : float ): return round(float((moles * 0.0_821 * temperature) / (pressure) ) ) def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float , __lowerCAmelCase : float ): return round(float((pressure * volume) / (0.0_821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
240
0
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch _snake_case = logging.get_logger(__name__) class _snake_case : def __init__( self: Union[str, Any] , __lowerCamelCase: str = None , __lowerCamelCase: uuid.UUID = None , __lowerCamelCase: Optional[Any]=None , __lowerCamelCase: Optional[int]=None ) -> Dict: if not conversation_id: __UpperCAmelCase : Tuple = uuid.uuida() if past_user_inputs is None: __UpperCAmelCase : Any = [] if generated_responses is None: __UpperCAmelCase : List[Any] = [] __UpperCAmelCase : uuid.UUID = conversation_id __UpperCAmelCase : List[str] = past_user_inputs __UpperCAmelCase : List[str] = generated_responses __UpperCAmelCase : Optional[str] = text def __eq__( self: Tuple , __lowerCamelCase: Dict ) -> List[str]: if not isinstance(__lowerCamelCase , __lowerCamelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def _lowerCamelCase ( self: Any , __lowerCamelCase: str , __lowerCamelCase: bool = False ) -> List[Any]: if self.new_user_input: if overwrite: logger.warning( f'''User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ''' f'''with: "{text}".''' ) __UpperCAmelCase : str = text else: logger.warning( f'''User input added while unprocessed input was existing: "{self.new_user_input}" new input ''' f'''ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input''' ) else: __UpperCAmelCase : int = text def _lowerCamelCase ( self: List[Any] ) -> Dict: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __UpperCAmelCase : Any = None def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: str ) -> int: self.generated_responses.append(__lowerCamelCase ) def _lowerCamelCase ( self: Optional[Any] ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self: Optional[int] ) -> int: __UpperCAmelCase : Tuple = f'''Conversation id: {self.uuid} \n''' for is_user, text in self.iter_texts(): __UpperCAmelCase : Optional[int] = "user" if is_user else "bot" output += f'''{name} >> {text} \n''' return output @add_end_docstrings( _lowercase , R"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class _snake_case ( _lowercase ): def __init__( self: str , *__lowerCamelCase: Union[str, Any] , **__lowerCamelCase: Dict ) -> Optional[int]: super().__init__(*__lowerCamelCase , **__lowerCamelCase ) if self.tokenizer.pad_token_id is None: __UpperCAmelCase : Any = self.tokenizer.eos_token def _lowerCamelCase ( self: List[str] , __lowerCamelCase: List[str]=None , __lowerCamelCase: Optional[int]=None , __lowerCamelCase: Tuple=None , **__lowerCamelCase: int ) -> int: __UpperCAmelCase : int = {} __UpperCAmelCase : Union[str, Any] = {} __UpperCAmelCase : Union[str, Any] = {} if min_length_for_response is not None: __UpperCAmelCase : Dict = min_length_for_response if minimum_tokens is not None: __UpperCAmelCase : str = minimum_tokens if "max_length" in generate_kwargs: __UpperCAmelCase : Any = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __UpperCAmelCase : Dict = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(__lowerCamelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self: Dict , __lowerCamelCase: Union[Conversation, List[Conversation]] , __lowerCamelCase: List[str]=0 , **__lowerCamelCase: str ) -> Optional[Any]: __UpperCAmelCase : Tuple = super().__call__(__lowerCamelCase , num_workers=__lowerCamelCase , **__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) and len(__lowerCamelCase ) == 1: return outputs[0] return outputs def _lowerCamelCase ( self: Dict , __lowerCamelCase: Conversation , __lowerCamelCase: str=32 ) -> Dict[str, Any]: if not isinstance(__lowerCamelCase , __lowerCamelCase ): raise ValueError("ConversationalPipeline, expects Conversation as inputs" ) if conversation.new_user_input is None: raise ValueError( f'''Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ''' "Add user inputs with the conversation's `add_user_input` method" ) if hasattr(self.tokenizer , "_build_conversation_input_ids" ): __UpperCAmelCase : Optional[int] = self.tokenizer._build_conversation_input_ids(__lowerCamelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version __UpperCAmelCase : Dict = self._legacy_parse_and_tokenize(__lowerCamelCase ) if self.framework == "pt": __UpperCAmelCase : List[Any] = torch.LongTensor([input_ids] ) elif self.framework == "tf": __UpperCAmelCase : List[Any] = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def _lowerCamelCase ( self: str , __lowerCamelCase: Optional[int] , __lowerCamelCase: List[Any]=10 , **__lowerCamelCase: str ) -> int: __UpperCAmelCase : str = generate_kwargs.get("max_length" , self.model.config.max_length ) __UpperCAmelCase : Any = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'''Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})''' ) __UpperCAmelCase : List[Any] = max_length - minimum_tokens __UpperCAmelCase : Union[str, Any] = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: __UpperCAmelCase : Any = model_inputs["attention_mask"][:, -trim:] __UpperCAmelCase : Dict = model_inputs.pop("conversation" ) __UpperCAmelCase : Tuple = max_length __UpperCAmelCase : Tuple = self.model.generate(**__lowerCamelCase , **__lowerCamelCase ) if self.model.config.is_encoder_decoder: __UpperCAmelCase : Optional[int] = 1 else: __UpperCAmelCase : Any = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def _lowerCamelCase ( self: Tuple , __lowerCamelCase: List[Any] , __lowerCamelCase: Dict=True ) -> List[Any]: __UpperCAmelCase : List[str] = model_outputs["output_ids"] __UpperCAmelCase : Tuple = self.tokenizer.decode( output_ids[0] , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase , ) __UpperCAmelCase : List[Any] = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(__lowerCamelCase ) return conversation def _lowerCamelCase ( self: Dict , __lowerCamelCase: Conversation ) -> Dict: __UpperCAmelCase : int = self.tokenizer.eos_token_id __UpperCAmelCase : int = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) ) if len(__lowerCamelCase ) > self.tokenizer.model_max_length: __UpperCAmelCase : Union[str, Any] = input_ids[-self.tokenizer.model_max_length :] return input_ids
342
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) def _UpperCamelCase ( snake_case__ ) -> int: __UpperCAmelCase : int = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: __UpperCAmelCase : int = [144, 192, 240] __UpperCAmelCase : Optional[Any] = [16, 32, 64, 96, 128, 160, 640] elif "mobilevit_xs" in mobilevit_name: __UpperCAmelCase : Optional[Any] = [96, 120, 144] __UpperCAmelCase : Tuple = [16, 32, 48, 64, 80, 96, 384] elif "mobilevit_xxs" in mobilevit_name: __UpperCAmelCase : str = [64, 80, 96] __UpperCAmelCase : Optional[Any] = [16, 16, 24, 48, 64, 80, 320] __UpperCAmelCase : Tuple = 0.05 __UpperCAmelCase : Dict = 2.0 if mobilevit_name.startswith("deeplabv3_" ): __UpperCAmelCase : str = 512 __UpperCAmelCase : Any = 16 __UpperCAmelCase : str = 21 __UpperCAmelCase : Union[str, Any] = "pascal-voc-id2label.json" else: __UpperCAmelCase : Optional[Any] = 1000 __UpperCAmelCase : int = "imagenet-1k-id2label.json" __UpperCAmelCase : Dict = "huggingface/label-files" __UpperCAmelCase : int = json.load(open(hf_hub_download(snake_case__, snake_case__, repo_type="dataset" ), "r" ) ) __UpperCAmelCase : Any = {int(snake_case__ ): v for k, v in idalabel.items()} __UpperCAmelCase : int = idalabel __UpperCAmelCase : List[str] = {v: k for k, v in idalabel.items()} return config def _UpperCamelCase ( snake_case__, snake_case__=False ) -> Tuple: for i in range(1, 6 ): if f'''layer_{i}.''' in name: __UpperCAmelCase : Tuple = name.replace(f'''layer_{i}.''', f'''encoder.layer.{i - 1}.''' ) if "conv_1." in name: __UpperCAmelCase : Dict = name.replace("conv_1.", "conv_stem." ) if ".block." in name: __UpperCAmelCase : Optional[int] = name.replace(".block.", "." ) if "exp_1x1" in name: __UpperCAmelCase : Tuple = name.replace("exp_1x1", "expand_1x1" ) if "red_1x1" in name: __UpperCAmelCase : Optional[Any] = name.replace("red_1x1", "reduce_1x1" ) if ".local_rep.conv_3x3." in name: __UpperCAmelCase : Optional[int] = name.replace(".local_rep.conv_3x3.", ".conv_kxk." ) if ".local_rep.conv_1x1." in name: __UpperCAmelCase : Any = name.replace(".local_rep.conv_1x1.", ".conv_1x1." ) if ".norm." in name: __UpperCAmelCase : Dict = name.replace(".norm.", ".normalization." ) if ".conv." in name: __UpperCAmelCase : List[Any] = name.replace(".conv.", ".convolution." ) if ".conv_proj." in name: __UpperCAmelCase : List[str] = name.replace(".conv_proj.", ".conv_projection." ) for i in range(0, 2 ): for j in range(0, 4 ): if f'''.{i}.{j}.''' in name: __UpperCAmelCase : List[Any] = name.replace(f'''.{i}.{j}.''', f'''.{i}.layer.{j}.''' ) for i in range(2, 6 ): for j in range(0, 4 ): if f'''.{i}.{j}.''' in name: __UpperCAmelCase : Any = name.replace(f'''.{i}.{j}.''', f'''.{i}.''' ) if "expand_1x1" in name: __UpperCAmelCase : Optional[int] = name.replace("expand_1x1", "downsampling_layer.expand_1x1" ) if "conv_3x3" in name: __UpperCAmelCase : List[Any] = name.replace("conv_3x3", "downsampling_layer.conv_3x3" ) if "reduce_1x1" in name: __UpperCAmelCase : Dict = name.replace("reduce_1x1", "downsampling_layer.reduce_1x1" ) for i in range(2, 5 ): if f'''.global_rep.{i}.weight''' in name: __UpperCAmelCase : Any = name.replace(f'''.global_rep.{i}.weight''', ".layernorm.weight" ) if f'''.global_rep.{i}.bias''' in name: __UpperCAmelCase : Optional[Any] = name.replace(f'''.global_rep.{i}.bias''', ".layernorm.bias" ) if ".global_rep." in name: __UpperCAmelCase : Tuple = name.replace(".global_rep.", ".transformer." ) if ".pre_norm_mha.0." in name: __UpperCAmelCase : Optional[Any] = name.replace(".pre_norm_mha.0.", ".layernorm_before." ) if ".pre_norm_mha.1.out_proj." in name: __UpperCAmelCase : Tuple = name.replace(".pre_norm_mha.1.out_proj.", ".attention.output.dense." ) if ".pre_norm_ffn.0." in name: __UpperCAmelCase : Optional[Any] = name.replace(".pre_norm_ffn.0.", ".layernorm_after." ) if ".pre_norm_ffn.1." in name: __UpperCAmelCase : Dict = name.replace(".pre_norm_ffn.1.", ".intermediate.dense." ) if ".pre_norm_ffn.4." in name: __UpperCAmelCase : int = name.replace(".pre_norm_ffn.4.", ".output.dense." ) if ".transformer." in name: __UpperCAmelCase : Tuple = name.replace(".transformer.", ".transformer.layer." ) if ".aspp_layer." in name: __UpperCAmelCase : Any = name.replace(".aspp_layer.", "." ) if ".aspp_pool." in name: __UpperCAmelCase : Optional[Any] = name.replace(".aspp_pool.", "." ) if "seg_head." in name: __UpperCAmelCase : Optional[int] = name.replace("seg_head.", "segmentation_head." ) if "segmentation_head.classifier.classifier." in name: __UpperCAmelCase : str = name.replace("segmentation_head.classifier.classifier.", "segmentation_head.classifier." ) if "classifier.fc." in name: __UpperCAmelCase : Optional[Any] = name.replace("classifier.fc.", "classifier." ) elif (not base_model) and ("segmentation_head." not in name): __UpperCAmelCase : List[str] = "mobilevit." + name return name def _UpperCamelCase ( snake_case__, snake_case__, snake_case__=False ) -> Union[str, Any]: if base_model: __UpperCAmelCase : Optional[int] = "" else: __UpperCAmelCase : Tuple = "mobilevit." for key in orig_state_dict.copy().keys(): __UpperCAmelCase : Optional[int] = orig_state_dict.pop(snake_case__ ) if key[:8] == "encoder.": __UpperCAmelCase : str = key[8:] if "qkv" in key: __UpperCAmelCase : Tuple = key.split("." ) __UpperCAmelCase : List[Any] = int(key_split[0][6:] ) - 1 __UpperCAmelCase : Optional[Any] = int(key_split[3] ) __UpperCAmelCase : Tuple = model.get_submodule(f'''{model_prefix}encoder.layer.{layer_num}''' ) __UpperCAmelCase : List[str] = layer.transformer.layer[transformer_num].attention.attention.all_head_size __UpperCAmelCase : Optional[Any] = ( f'''{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.''' ) if "weight" in key: __UpperCAmelCase : Any = val[:dim, :] __UpperCAmelCase : Any = val[dim : dim * 2, :] __UpperCAmelCase : List[Any] = val[-dim:, :] else: __UpperCAmelCase : List[str] = val[:dim] __UpperCAmelCase : Optional[Any] = val[dim : dim * 2] __UpperCAmelCase : List[Any] = val[-dim:] else: __UpperCAmelCase : str = val return orig_state_dict def _UpperCamelCase ( ) -> Any: __UpperCAmelCase : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" __UpperCAmelCase : List[str] = Image.open(requests.get(snake_case__, stream=snake_case__ ).raw ) return im @torch.no_grad() def _UpperCamelCase ( snake_case__, snake_case__, snake_case__, snake_case__=False ) -> Optional[Any]: __UpperCAmelCase : Tuple = get_mobilevit_config(snake_case__ ) # load original state_dict __UpperCAmelCase : str = torch.load(snake_case__, map_location="cpu" ) # load 🤗 model if mobilevit_name.startswith("deeplabv3_" ): __UpperCAmelCase : Optional[int] = MobileViTForSemanticSegmentation(snake_case__ ).eval() else: __UpperCAmelCase : List[Any] = MobileViTForImageClassification(snake_case__ ).eval() __UpperCAmelCase : Dict = convert_state_dict(snake_case__, snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by MobileViTImageProcessor __UpperCAmelCase : Optional[Any] = MobileViTImageProcessor(crop_size=config.image_size, size=config.image_size + 32 ) __UpperCAmelCase : Any = image_processor(images=prepare_img(), return_tensors="pt" ) __UpperCAmelCase : Dict = model(**snake_case__ ) __UpperCAmelCase : Tuple = outputs.logits if mobilevit_name.startswith("deeplabv3_" ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": __UpperCAmelCase : int = torch.tensor( [ [[6.2065, 6.1292, 6.2070], [6.1079, 6.1254, 6.1747], [6.0042, 6.1071, 6.1034]], [[-6.9253, -6.8653, -7.0398], [-7.3218, -7.3983, -7.3670], [-7.1961, -7.2482, -7.1569]], [[-4.4723, -4.4348, -4.3769], [-5.3629, -5.4632, -5.4598], [-5.1587, -5.3402, -5.5059]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": __UpperCAmelCase : Tuple = torch.tensor( [ [[5.4449, 5.5733, 5.6314], [5.1815, 5.3930, 5.5963], [5.1656, 5.4333, 5.4853]], [[-9.4423, -9.7766, -9.6714], [-9.1581, -9.5720, -9.5519], [-9.1006, -9.6458, -9.5703]], [[-7.7721, -7.3716, -7.1583], [-8.4599, -8.0624, -7.7944], [-8.4172, -7.8366, -7.5025]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": __UpperCAmelCase : Any = torch.tensor( [ [[6.9811, 6.9743, 7.3123], [7.1777, 7.1931, 7.3938], [7.5633, 7.8050, 7.8901]], [[-10.5536, -10.2332, -10.2924], [-10.2336, -9.8624, -9.5964], [-10.8840, -10.8158, -10.6659]], [[-3.4938, -3.0631, -2.8620], [-3.4205, -2.8135, -2.6875], [-3.4179, -2.7945, -2.8750]], ] ) else: raise ValueError(f'''Unknown mobilevit_name: {mobilevit_name}''' ) assert torch.allclose(logits[0, :3, :3, :3], snake_case__, atol=1e-4 ) else: assert logits.shape == (1, 1000) if mobilevit_name == "mobilevit_s": __UpperCAmelCase : str = torch.tensor([-0.9866, 0.2392, -1.1241] ) elif mobilevit_name == "mobilevit_xs": __UpperCAmelCase : Tuple = torch.tensor([-2.4761, -0.9399, -1.9587] ) elif mobilevit_name == "mobilevit_xxs": __UpperCAmelCase : Union[str, Any] = torch.tensor([-1.9364, -1.2327, -0.4653] ) else: raise ValueError(f'''Unknown mobilevit_name: {mobilevit_name}''' ) assert torch.allclose(logits[0, :3], snake_case__, atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(f'''Saving model {mobilevit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(snake_case__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: __UpperCAmelCase : List[str] = { "mobilevit_s": "mobilevit-small", "mobilevit_xs": "mobilevit-x-small", "mobilevit_xxs": "mobilevit-xx-small", "deeplabv3_mobilevit_s": "deeplabv3-mobilevit-small", "deeplabv3_mobilevit_xs": "deeplabv3-mobilevit-x-small", "deeplabv3_mobilevit_xxs": "deeplabv3-mobilevit-xx-small", } print("Pushing to the hub..." ) __UpperCAmelCase : int = model_mapping[mobilevit_name] image_processor.push_to_hub(snake_case__, organization="apple" ) model.push_to_hub(snake_case__, organization="apple" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--mobilevit_name''', default='''mobilevit_s''', type=str, help=( '''Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\',''' ''' \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'.''' ), ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original state dict (.pt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _snake_case = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
342
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase : Optional[int] = logging.get_logger(__name__) lowerCamelCase : Any = { '''microsoft/beit-base-patch16-224-pt22k''': ( '''https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json''' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class lowerCAmelCase ( __a ): '''simple docstring''' _A : Dict = '''beit''' def __init__( self : List[Any] , __a : Tuple=8192 , __a : List[Any]=768 , __a : Optional[int]=12 , __a : List[str]=12 , __a : Optional[int]=3072 , __a : str="gelu" , __a : Union[str, Any]=0.0 , __a : Optional[Any]=0.0 , __a : List[str]=0.02 , __a : List[Any]=1E-12 , __a : Union[str, Any]=224 , __a : Optional[Any]=16 , __a : List[Any]=3 , __a : int=False , __a : Dict=False , __a : int=False , __a : str=False , __a : Any=0.1 , __a : int=0.1 , __a : List[Any]=True , __a : str=[3, 5, 7, 11] , __a : Dict=[1, 2, 3, 6] , __a : int=True , __a : Dict=0.4 , __a : List[str]=256 , __a : Dict=1 , __a : Dict=False , __a : str=255 , **__a : str , ) -> List[str]: """simple docstring""" super().__init__(**__a ) __lowercase : Any = vocab_size __lowercase : Any = hidden_size __lowercase : Any = num_hidden_layers __lowercase : Dict = num_attention_heads __lowercase : Dict = intermediate_size __lowercase : List[Any] = hidden_act __lowercase : Any = hidden_dropout_prob __lowercase : List[str] = attention_probs_dropout_prob __lowercase : List[str] = initializer_range __lowercase : str = layer_norm_eps __lowercase : List[str] = image_size __lowercase : Optional[int] = patch_size __lowercase : Optional[int] = num_channels __lowercase : Tuple = use_mask_token __lowercase : Dict = use_absolute_position_embeddings __lowercase : str = use_relative_position_bias __lowercase : List[Any] = use_shared_relative_position_bias __lowercase : Union[str, Any] = layer_scale_init_value __lowercase : Tuple = drop_path_rate __lowercase : Union[str, Any] = use_mean_pooling # decode head attributes (semantic segmentation) __lowercase : Dict = out_indices __lowercase : Dict = pool_scales # auxiliary head attributes (semantic segmentation) __lowercase : Optional[int] = use_auxiliary_head __lowercase : List[str] = auxiliary_loss_weight __lowercase : Optional[int] = auxiliary_channels __lowercase : Tuple = auxiliary_num_convs __lowercase : Optional[int] = auxiliary_concat_input __lowercase : int = semantic_loss_ignore_index class lowerCAmelCase ( __a ): '''simple docstring''' _A : int = version.parse('''1.11''' ) @property def lowerCAmelCase ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCAmelCase ( self : Optional[Any] ) -> float: """simple docstring""" return 1E-4
233
import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase : Any = get_tests_dir('''fixtures/test_sentencepiece.model''') lowerCamelCase : Any = get_tests_dir('''fixtures/test_sentencepiece_bpe.model''') lowerCamelCase : Optional[int] = '''pt''' if is_torch_available() else '''tf''' @require_sentencepiece @require_tokenizers class lowerCAmelCase ( __a , unittest.TestCase ): '''simple docstring''' _A : Dict = CamembertTokenizer _A : Union[str, Any] = CamembertTokenizerFast _A : Union[str, Any] = True _A : Tuple = True def lowerCAmelCase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __lowercase : Union[str, Any] = CamembertTokenizer(__a ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" __lowercase : Dict = """<pad>""" __lowercase : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__a ) , __a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__a ) , __a ) def lowerCAmelCase ( self : Optional[Any] ) -> Tuple: """simple docstring""" __lowercase : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>NOTUSED""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__a ) , 1004 ) def lowerCAmelCase ( self : str ) -> List[Any]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1005 ) def lowerCAmelCase ( self : Any ) -> Optional[Any]: """simple docstring""" __lowercase : Tuple = CamembertTokenizer(__a ) tokenizer.save_pretrained(self.tmpdirname ) __lowercase : Tuple = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) __lowercase : List[str] = """I was born in 92000, and this is falsé.""" __lowercase : Optional[Any] = tokenizer.encode(__a ) __lowercase : List[Any] = rust_tokenizer.encode(__a ) self.assertListEqual(__a , __a ) __lowercase : Tuple = tokenizer.encode(__a , add_special_tokens=__a ) __lowercase : Union[str, Any] = rust_tokenizer.encode(__a , add_special_tokens=__a ) self.assertListEqual(__a , __a ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) __lowercase : Dict = tokenizer.convert_ids_to_tokens(__a ) __lowercase : Any = rust_tokenizer.tokenize(__a ) self.assertListEqual(__a , __a ) def lowerCAmelCase ( self : Any ) -> Dict: """simple docstring""" if not self.test_rust_tokenizer: return __lowercase : List[str] = self.get_tokenizer() __lowercase : Any = self.get_rust_tokenizer() __lowercase : Any = """I was born in 92000, and this is falsé.""" __lowercase : Tuple = tokenizer.tokenize(__a ) __lowercase : Optional[Any] = rust_tokenizer.tokenize(__a ) self.assertListEqual(__a , __a ) __lowercase : Dict = tokenizer.encode(__a , add_special_tokens=__a ) __lowercase : Dict = rust_tokenizer.encode(__a , add_special_tokens=__a ) self.assertListEqual(__a , __a ) __lowercase : Any = self.get_rust_tokenizer() __lowercase : str = tokenizer.encode(__a ) __lowercase : Union[str, Any] = rust_tokenizer.encode(__a ) self.assertListEqual(__a , __a ) @slow def lowerCAmelCase ( self : Optional[int] ) -> List[Any]: """simple docstring""" __lowercase : str = {"""input_ids""": [[5, 54, 7196, 297, 30, 23, 776, 18, 11, 3215, 3705, 8252, 22, 3164, 1181, 2116, 29, 16, 813, 25, 791, 3314, 20, 3446, 38, 27575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9088, 20, 1517, 8, 22804, 18818, 10, 38, 629, 607, 607, 142, 19, 7196, 867, 56, 10326, 24, 2267, 20, 416, 5072, 15612, 233, 734, 7, 2399, 27, 16, 3015, 1649, 7, 24, 20, 4338, 2399, 27, 13, 3400, 14, 13, 6189, 8, 930, 9, 6]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. __lowercase : List[str] = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=__a , model_name="""camembert-base""" , revision="""3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf""" , sequences=__a , )
233
1
def A ( __UpperCAmelCase , __UpperCAmelCase = False ) -> Optional[Any]: '''simple docstring''' if not isinstance(__a , __a ): UpperCAmelCase_ = f"Expected string as input, found {type(__a )}" raise ValueError(__a ) if not isinstance(__a , __a ): UpperCAmelCase_ = f"Expected boolean as use_pascal parameter, found {type(__a )}" raise ValueError(__a ) UpperCAmelCase_ = input_str.split('''_''' ) UpperCAmelCase_ = 0 if use_pascal else 1 UpperCAmelCase_ = words[start_index:] UpperCAmelCase_ = [word[0].upper() + word[1:] for word in words_to_capitalize] UpperCAmelCase_ = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
350
def A ( __UpperCAmelCase = 100_0000 ) -> int: '''simple docstring''' UpperCAmelCase_ = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __UpperCAmelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
344
0
'''simple docstring''' # XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path __UpperCAmelCase =Path(__file__).resolve().parents[3] / "src" sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(4_2) __UpperCAmelCase ={"base": "patrickvonplaten/wav2vec2_tiny_random", "robust": "patrickvonplaten/wav2vec2_tiny_random_robust"} __UpperCAmelCase ="zero2" __UpperCAmelCase ="zero3" __UpperCAmelCase =[ZEROa, ZEROa] def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Any: __lowerCamelCase = parameterized.to_safe_name('''_'''.join(str(_A ) for x in param.args ) ) return f"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test __UpperCAmelCase =list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class a__ ( lowerCamelCase__ ): @parameterized.expand(lowercase_ , name_func=lowercase_ ) def SCREAMING_SNAKE_CASE__ ( self : Dict , a : Union[str, Any] , a : Tuple ): """simple docstring""" self.run_and_check( stage=lowercase_ , model=lowercase_ , distributed=lowercase_ , fpaa=lowercase_ , ) @require_torch_multi_gpu @parameterized.expand(lowercase_ , name_func=lowercase_ ) def SCREAMING_SNAKE_CASE__ ( self : Any , a : Tuple , a : Optional[int] ): """simple docstring""" self.run_and_check( stage=lowercase_ , model=lowercase_ , distributed=lowercase_ , fpaa=lowercase_ , ) @parameterized.expand(lowercase_ , name_func=lowercase_ ) def SCREAMING_SNAKE_CASE__ ( self : Tuple , a : Any , a : Optional[Any] ): """simple docstring""" self.run_and_check( stage=lowercase_ , model=lowercase_ , distributed=lowercase_ , fpaa=lowercase_ , ) @require_torch_multi_gpu @parameterized.expand(lowercase_ , name_func=lowercase_ ) def SCREAMING_SNAKE_CASE__ ( self : int , a : Any , a : Dict ): """simple docstring""" self.run_and_check( stage=lowercase_ , model=lowercase_ , distributed=lowercase_ , fpaa=lowercase_ , ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , a : List[str] ): """simple docstring""" pass def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : int , a : Dict , a : int = 10 , a : List[str] = True , a : Union[str, Any] = True , a : str = True , ): """simple docstring""" __lowerCamelCase = models[model] __lowerCamelCase = self.run_trainer( stage=lowercase_ , model_name=lowercase_ , eval_steps=lowercase_ , num_train_epochs=1 , distributed=lowercase_ , fpaa=lowercase_ , ) self.do_checks(lowercase_ ) return output_dir def SCREAMING_SNAKE_CASE__ ( self : Tuple , a : Dict , a : Optional[Any] , a : str = 10 , a : Tuple = 1 , a : int = True , a : Optional[int] = True , ): """simple docstring""" __lowerCamelCase = self.get_auto_remove_tmp_dir('''./xxx''' , after=lowercase_ ) __lowerCamelCase = f""" --model_name_or_path {model_name} --dataset_name hf-internal-testing/librispeech_asr_dummy --dataset_config_name clean --train_split_name validation --validation_split_name validation --output_dir {output_dir} --num_train_epochs {str(lowercase_ )} --per_device_train_batch_size 2 --per_device_eval_batch_size 2 --evaluation_strategy steps --learning_rate 5e-4 --warmup_steps 8 --orthography timit --preprocessing_num_workers 1 --group_by_length --freeze_feature_extractor --report_to none --save_steps 0 --eval_steps {eval_steps} --report_to none """.split() if fpaa: args.extend(['''--fp16'''] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files __lowerCamelCase = f"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() __lowerCamelCase = [f"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] __lowerCamelCase = self.get_launcher(lowercase_ ) __lowerCamelCase = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowercase_ , env=self.get_env() ) return output_dir def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : Tuple=False ): """simple docstring""" __lowerCamelCase = min(2 , get_gpu_count() ) if distributed else 1 return f"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
67
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
188
0
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor lowerCamelCase_ = logging.get_logger(__name__) class __lowerCamelCase ( __snake_case ): def __init__( self , *lowerCamelCase , **lowerCamelCase ) -> None: warnings.warn( """The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use BeitImageProcessor instead.""" , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
34
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCamelCase_ = { '''vocab_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json''' ), }, } lowerCamelCase_ = { '''yjernite/retribert-base-uncased''': 512, } lowerCamelCase_ = { '''yjernite/retribert-base-uncased''': {'''do_lower_case''': True}, } class __lowerCamelCase ( __snake_case ): lowerCamelCase_ : Union[str, Any] = VOCAB_FILES_NAMES lowerCamelCase_ : str = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ : Optional[int] = PRETRAINED_INIT_CONFIGURATION lowerCamelCase_ : Union[str, Any] = RetriBertTokenizer lowerCamelCase_ : str = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase="[UNK]" , lowerCamelCase="[SEP]" , lowerCamelCase="[PAD]" , lowerCamelCase="[CLS]" , lowerCamelCase="[MASK]" , lowerCamelCase=True , lowerCamelCase=None , **lowerCamelCase , ) -> List[Any]: super().__init__( lowerCamelCase , tokenizer_file=lowerCamelCase , do_lower_case=lowerCamelCase , unk_token=lowerCamelCase , sep_token=lowerCamelCase , pad_token=lowerCamelCase , cls_token=lowerCamelCase , mask_token=lowerCamelCase , tokenize_chinese_chars=lowerCamelCase , strip_accents=lowerCamelCase , **lowerCamelCase , ) snake_case_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , lowerCamelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , lowerCamelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , lowerCamelCase ) != tokenize_chinese_chars ): snake_case_ = getattr(lowerCamelCase , normalizer_state.pop("""type""" ) ) snake_case_ = do_lower_case snake_case_ = strip_accents snake_case_ = tokenize_chinese_chars snake_case_ = normalizer_class(**lowerCamelCase ) snake_case_ = do_lower_case def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase=None ) -> str: snake_case_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase = None ) -> List[int]: snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase = None ) -> Tuple[str]: snake_case_ = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase )
34
1
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from utils_ner import Split, TokenClassificationDataset, TokenClassificationTask import transformers from transformers import ( AutoConfig, AutoModelForTokenClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process UpperCAmelCase__ = logging.getLogger(__name__) @dataclass class a : _snake_case : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) _snake_case : Optional[str] = field( default=a__ , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) _snake_case : Optional[str] = field( default='NER' , metadata={'help': 'Task type to fine tune in training (e.g. NER, POS, etc)'} ) _snake_case : Optional[str] = field( default=a__ , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) _snake_case : bool = field(default=a__ , metadata={'help': 'Set this flag to use fast tokenization.'} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. _snake_case : Optional[str] = field( default=a__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) @dataclass class a : _snake_case : str = field( metadata={'help': 'The input data dir. Should contain the .txt files for a CoNLL-2003-formatted task.'} ) _snake_case : Optional[str] = field( default=a__ , metadata={'help': 'Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.'} , ) _snake_case : int = field( default=1_28 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) _snake_case : bool = field( default=a__ , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) def __UpperCAmelCase ( ): """simple docstring""" # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' """ --overwrite_output_dir to overcome.""" ) _UpperCAmelCase = import_module("""tasks""" ) try: _UpperCAmelCase = getattr(_snake_case ,model_args.task_type ) _UpperCAmelCase = token_classification_task_clazz() except AttributeError: raise ValueError( f'''Task {model_args.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' f'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" ,datefmt="""%m/%d/%Y %H:%M:%S""" ,level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN ,) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" ,training_args.local_rank ,training_args.device ,training_args.n_gpu ,bool(training_args.local_rank != -1 ) ,training_args.fpaa ,) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" ,_snake_case ) # Set seed set_seed(training_args.seed ) # Prepare CONLL-2003 task _UpperCAmelCase = token_classification_task.get_labels(data_args.labels ) _UpperCAmelCase = dict(enumerate(_snake_case ) ) _UpperCAmelCase = len(_snake_case ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _UpperCAmelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path ,num_labels=_snake_case ,idalabel=_snake_case ,labelaid={label: i for i, label in enumerate(_snake_case )} ,cache_dir=model_args.cache_dir ,) _UpperCAmelCase = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path ,cache_dir=model_args.cache_dir ,use_fast=model_args.use_fast ,) _UpperCAmelCase = AutoModelForTokenClassification.from_pretrained( model_args.model_name_or_path ,from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) ,config=_snake_case ,cache_dir=model_args.cache_dir ,) # Get datasets _UpperCAmelCase = ( TokenClassificationDataset( token_classification_task=_snake_case ,data_dir=data_args.data_dir ,tokenizer=_snake_case ,labels=_snake_case ,model_type=config.model_type ,max_seq_length=data_args.max_seq_length ,overwrite_cache=data_args.overwrite_cache ,mode=Split.train ,) if training_args.do_train else None ) _UpperCAmelCase = ( TokenClassificationDataset( token_classification_task=_snake_case ,data_dir=data_args.data_dir ,tokenizer=_snake_case ,labels=_snake_case ,model_type=config.model_type ,max_seq_length=data_args.max_seq_length ,overwrite_cache=data_args.overwrite_cache ,mode=Split.dev ,) if training_args.do_eval else None ) def align_predictions(lowercase ,lowercase ) -> Tuple[List[int], List[int]]: _UpperCAmelCase = np.argmax(_snake_case ,axis=2 ) _UpperCAmelCase = preds.shape _UpperCAmelCase = [[] for _ in range(_snake_case )] _UpperCAmelCase = [[] for _ in range(_snake_case )] for i in range(_snake_case ): for j in range(_snake_case ): if label_ids[i, j] != nn.CrossEntropyLoss().ignore_index: out_label_list[i].append(label_map[label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) return preds_list, out_label_list def compute_metrics(lowercase ) -> Dict: _UpperCAmelCase = align_predictions(p.predictions ,p.label_ids ) return { "accuracy_score": accuracy_score(_snake_case ,_snake_case ), "precision": precision_score(_snake_case ,_snake_case ), "recall": recall_score(_snake_case ,_snake_case ), "f1": fa_score(_snake_case ,_snake_case ), } # Data collator _UpperCAmelCase = DataCollatorWithPadding(_snake_case ,pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer _UpperCAmelCase = Trainer( model=_snake_case ,args=_snake_case ,train_dataset=_snake_case ,eval_dataset=_snake_case ,compute_metrics=_snake_case ,data_collator=_snake_case ,) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_process_zero(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation _UpperCAmelCase = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) _UpperCAmelCase = trainer.evaluate() _UpperCAmelCase = os.path.join(training_args.output_dir ,"""eval_results.txt""" ) if trainer.is_world_process_zero(): with open(_snake_case ,"""w""" ) as writer: logger.info("""***** Eval results *****""" ) for key, value in result.items(): logger.info(""" %s = %s""" ,_snake_case ,_snake_case ) writer.write("""%s = %s\n""" % (key, value) ) results.update(_snake_case ) # Predict if training_args.do_predict: _UpperCAmelCase = TokenClassificationDataset( token_classification_task=_snake_case ,data_dir=data_args.data_dir ,tokenizer=_snake_case ,labels=_snake_case ,model_type=config.model_type ,max_seq_length=data_args.max_seq_length ,overwrite_cache=data_args.overwrite_cache ,mode=Split.test ,) _UpperCAmelCase = trainer.predict(_snake_case ) _UpperCAmelCase = align_predictions(_snake_case ,_snake_case ) _UpperCAmelCase = os.path.join(training_args.output_dir ,"""test_results.txt""" ) if trainer.is_world_process_zero(): with open(_snake_case ,"""w""" ) as writer: for key, value in metrics.items(): logger.info(""" %s = %s""" ,_snake_case ,_snake_case ) writer.write("""%s = %s\n""" % (key, value) ) # Save predictions _UpperCAmelCase = os.path.join(training_args.output_dir ,"""test_predictions.txt""" ) if trainer.is_world_process_zero(): with open(_snake_case ,"""w""" ) as writer: with open(os.path.join(data_args.data_dir ,"""test.txt""" ) ,"""r""" ) as f: token_classification_task.write_predictions_to_file(_snake_case ,_snake_case ,_snake_case ) return results def __UpperCAmelCase ( lowercase ): """simple docstring""" # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
289
"""simple docstring""" import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('1.0.0a'): raise Exception('requires fairseq >= 1.0.0a') logging.set_verbosity_info() UpperCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) UpperCAmelCase__ : List[str] = 'Hello world! cécé herlolip' def lowercase_ ( _snake_case ,_snake_case ,_snake_case ): SCREAMING_SNAKE_CASE__ : int = FairseqRobertaModel.from_pretrained(_snake_case ) roberta.eval() # disable dropout SCREAMING_SNAKE_CASE__ : Any = roberta.model.encoder.sentence_encoder SCREAMING_SNAKE_CASE__ : Any = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings ,hidden_size=roberta.cfg.model.encoder_embed_dim ,num_hidden_layers=roberta.cfg.model.encoder_layers ,num_attention_heads=roberta.cfg.model.encoder_attention_heads ,intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim ,max_position_embeddings=514 ,type_vocab_size=1 ,layer_norm_eps=1E-5 ,) if classification_head: SCREAMING_SNAKE_CASE__ : Dict = roberta.model.classification_heads["""mnli"""].out_proj.weight.shape[0] print("""Our RoBERTa config:""" ,_snake_case ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = XLMRobertaXLForSequenceClassification(_snake_case ) if classification_head else XLMRobertaXLForMaskedLM(_snake_case ) model.eval() # Now let's copy all the weights. # Embeddings SCREAMING_SNAKE_CASE__ : Optional[int] = roberta_sent_encoder.embed_tokens.weight SCREAMING_SNAKE_CASE__ : int = roberta_sent_encoder.embed_positions.weight SCREAMING_SNAKE_CASE__ : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. SCREAMING_SNAKE_CASE__ : Optional[int] = roberta_sent_encoder.layer_norm.weight SCREAMING_SNAKE_CASE__ : Dict = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer SCREAMING_SNAKE_CASE__ : BertLayer = model.roberta.encoder.layer[i] SCREAMING_SNAKE_CASE__ : TransformerSentenceEncoderLayer = roberta_sent_encoder.layers[i] SCREAMING_SNAKE_CASE__ : RobertaAttention = layer.attention SCREAMING_SNAKE_CASE__ : List[str] = roberta_layer.self_attn_layer_norm.weight SCREAMING_SNAKE_CASE__ : List[Any] = roberta_layer.self_attn_layer_norm.bias # self attention SCREAMING_SNAKE_CASE__ : BertSelfAttention = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) SCREAMING_SNAKE_CASE__ : List[str] = roberta_layer.self_attn.q_proj.weight SCREAMING_SNAKE_CASE__ : Tuple = roberta_layer.self_attn.q_proj.bias SCREAMING_SNAKE_CASE__ : Tuple = roberta_layer.self_attn.k_proj.weight SCREAMING_SNAKE_CASE__ : int = roberta_layer.self_attn.k_proj.bias SCREAMING_SNAKE_CASE__ : List[str] = roberta_layer.self_attn.v_proj.weight SCREAMING_SNAKE_CASE__ : Union[str, Any] = roberta_layer.self_attn.v_proj.bias # self-attention output SCREAMING_SNAKE_CASE__ : BertSelfOutput = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape SCREAMING_SNAKE_CASE__ : Union[str, Any] = roberta_layer.self_attn.out_proj.weight SCREAMING_SNAKE_CASE__ : List[str] = roberta_layer.self_attn.out_proj.bias # this one is final layer norm SCREAMING_SNAKE_CASE__ : Tuple = roberta_layer.final_layer_norm.weight SCREAMING_SNAKE_CASE__ : Optional[int] = roberta_layer.final_layer_norm.bias # intermediate SCREAMING_SNAKE_CASE__ : BertIntermediate = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape SCREAMING_SNAKE_CASE__ : List[Any] = roberta_layer.fca.weight SCREAMING_SNAKE_CASE__ : Tuple = roberta_layer.fca.bias # output SCREAMING_SNAKE_CASE__ : BertOutput = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape SCREAMING_SNAKE_CASE__ : Tuple = roberta_layer.fca.weight SCREAMING_SNAKE_CASE__ : Optional[int] = roberta_layer.fca.bias # end of layer if classification_head: SCREAMING_SNAKE_CASE__ : List[Any] = roberta.model.classification_heads["""mnli"""].dense.weight SCREAMING_SNAKE_CASE__ : Optional[Any] = roberta.model.classification_heads["""mnli"""].dense.bias SCREAMING_SNAKE_CASE__ : Optional[Any] = roberta.model.classification_heads["""mnli"""].out_proj.weight SCREAMING_SNAKE_CASE__ : Union[str, Any] = roberta.model.classification_heads["""mnli"""].out_proj.bias else: # LM Head SCREAMING_SNAKE_CASE__ : str = roberta.model.encoder.lm_head.dense.weight SCREAMING_SNAKE_CASE__ : List[Any] = roberta.model.encoder.lm_head.dense.bias SCREAMING_SNAKE_CASE__ : Union[str, Any] = roberta.model.encoder.lm_head.layer_norm.weight SCREAMING_SNAKE_CASE__ : Dict = roberta.model.encoder.lm_head.layer_norm.bias SCREAMING_SNAKE_CASE__ : Optional[int] = roberta.model.encoder.lm_head.weight SCREAMING_SNAKE_CASE__ : List[str] = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. SCREAMING_SNAKE_CASE__ : torch.Tensor = roberta.encode(_snake_case ).unsqueeze(0 ) # batch of size 1 SCREAMING_SNAKE_CASE__ : Tuple = model(_snake_case )[0] if classification_head: SCREAMING_SNAKE_CASE__ : Dict = roberta.model.classification_heads["""mnli"""](roberta.extract_features(_snake_case ) ) else: SCREAMING_SNAKE_CASE__ : Tuple = roberta.model(_snake_case )[0] print(our_output.shape ,their_output.shape ) SCREAMING_SNAKE_CASE__ : Optional[Any] = torch.max(torch.abs(our_output - their_output ) ).item() print(f'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 SCREAMING_SNAKE_CASE__ : Tuple = torch.allclose(_snake_case ,_snake_case ,atol=1E-3 ) print("""Do both models output the same tensors?""" ,"""🔥""" if success else """💩""" ) if not success: raise Exception("""Something went wRoNg""" ) pathlib.Path(_snake_case ).mkdir(parents=_snake_case ,exist_ok=_snake_case ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) if __name__ == "__main__": UpperCAmelCase__ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--roberta_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--classification_head', action='store_true', help='Whether to convert a final classification head.' ) UpperCAmelCase__ : Any = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
25
0
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer __lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Optional[Any] = 'AutoTokenizer' _UpperCAmelCase :Union[str, Any] = ['tokenizer'] _UpperCAmelCase :List[str] = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self , A_ , A_=None ): '''simple docstring''' super().__init__(A_ ) UpperCamelCase : Optional[Any] = speaker_embeddings @classmethod def __UpperCamelCase( cls , A_ , A_="speaker_embeddings_path.json" , **A_ ): '''simple docstring''' if speaker_embeddings_dict_path is not None: UpperCamelCase : Tuple = get_file_from_repo( A_ , A_ , subfolder=kwargs.pop("subfolder" , A_ ) , cache_dir=kwargs.pop("cache_dir" , A_ ) , force_download=kwargs.pop("force_download" , A_ ) , proxies=kwargs.pop("proxies" , A_ ) , resume_download=kwargs.pop("resume_download" , A_ ) , local_files_only=kwargs.pop("local_files_only" , A_ ) , use_auth_token=kwargs.pop("use_auth_token" , A_ ) , revision=kwargs.pop("revision" , A_ ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(A_ , A_ )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) UpperCamelCase : int = None else: with open(A_ ) as speaker_embeddings_json: UpperCamelCase : List[str] = json.load(A_ ) else: UpperCamelCase : Tuple = None UpperCamelCase : int = AutoTokenizer.from_pretrained(A_ , **A_ ) return cls(tokenizer=A_ , speaker_embeddings=A_ ) def __UpperCamelCase( self , A_ , A_="speaker_embeddings_path.json" , A_="speaker_embeddings" , A_ = False , **A_ , ): '''simple docstring''' if self.speaker_embeddings is not None: os.makedirs(os.path.join(A_ , A_ , "v2" ) , exist_ok=A_ ) UpperCamelCase : List[Any] = {} UpperCamelCase : Union[str, Any] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": UpperCamelCase : Union[str, Any] = self._load_voice_preset(A_ ) UpperCamelCase : Union[str, Any] = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , A_ , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=A_ , ) UpperCamelCase : Any = os.path.join(A_ , F"""{prompt_key}_{key}.npy""" ) UpperCamelCase : int = tmp_dict with open(os.path.join(A_ , A_ ) , "w" ) as fp: json.dump(A_ , A_ ) super().save_pretrained(A_ , A_ , **A_ ) def __UpperCamelCase( self , A_ = None , **A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.speaker_embeddings[voice_preset] UpperCamelCase : Dict = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) UpperCamelCase : int = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , A_ ) , cache_dir=kwargs.pop("cache_dir" , A_ ) , force_download=kwargs.pop("force_download" , A_ ) , proxies=kwargs.pop("proxies" , A_ ) , resume_download=kwargs.pop("resume_download" , A_ ) , local_files_only=kwargs.pop("local_files_only" , A_ ) , use_auth_token=kwargs.pop("use_auth_token" , A_ ) , revision=kwargs.pop("revision" , A_ ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) UpperCamelCase : str = np.load(A_ ) return voice_preset_dict def __UpperCamelCase( self , A_ = None ): '''simple docstring''' for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , A_=None , A_=None , A_="pt" , A_=256 , A_=False , A_=True , A_=False , **A_ , ): '''simple docstring''' if voice_preset is not None and not isinstance(A_ , A_ ): if ( isinstance(A_ , A_ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): UpperCamelCase : str = self._load_voice_preset(A_ ) else: if isinstance(A_ , A_ ) and not voice_preset.endswith(".npz" ): UpperCamelCase : Dict = voice_preset + ".npz" UpperCamelCase : int = np.load(A_ ) if voice_preset is not None: self._validate_voice_preset_dict(A_ , **A_ ) UpperCamelCase : List[Any] = BatchFeature(data=A_ , tensor_type=A_ ) UpperCamelCase : int = self.tokenizer( A_ , return_tensors=A_ , padding="max_length" , max_length=A_ , return_attention_mask=A_ , return_token_type_ids=A_ , add_special_tokens=A_ , **A_ , ) if voice_preset is not None: UpperCamelCase : Optional[int] = voice_preset return encoded_text
140
from __future__ import annotations def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> list[str]: if partitions <= 0: raise ValueError("partitions must be a positive number!" ) if partitions > number_of_bytes: raise ValueError("partitions can not > number_of_bytes!" ) UpperCamelCase : str = number_of_bytes // partitions UpperCamelCase : List[Any] = [] for i in range(_lowerCAmelCase ): UpperCamelCase : Optional[Any] = i * bytes_per_partition + 1 UpperCamelCase : Any = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(F"""{start_bytes}-{end_bytes}""" ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
140
1
import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCAmelCase : str = logging.get_logger("""transformers.models.speecht5""") def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: hf_model.apply_weight_norm() lowerCamelCase__ : str = checkpoint['input_conv.weight_g'] lowerCamelCase__ : List[Any] = checkpoint['input_conv.weight_v'] lowerCamelCase__ : Union[str, Any] = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): lowerCamelCase__ : Any = checkpoint[F"""upsamples.{i}.1.weight_g"""] lowerCamelCase__ : Optional[Any] = checkpoint[F"""upsamples.{i}.1.weight_v"""] lowerCamelCase__ : int = checkpoint[F"""upsamples.{i}.1.bias"""] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): lowerCamelCase__ : str = checkpoint[F"""blocks.{i}.convs1.{j}.1.weight_g"""] lowerCamelCase__ : Optional[int] = checkpoint[F"""blocks.{i}.convs1.{j}.1.weight_v"""] lowerCamelCase__ : Optional[Any] = checkpoint[F"""blocks.{i}.convs1.{j}.1.bias"""] lowerCamelCase__ : Tuple = checkpoint[F"""blocks.{i}.convs2.{j}.1.weight_g"""] lowerCamelCase__ : str = checkpoint[F"""blocks.{i}.convs2.{j}.1.weight_v"""] lowerCamelCase__ : List[Any] = checkpoint[F"""blocks.{i}.convs2.{j}.1.bias"""] lowerCamelCase__ : Tuple = checkpoint['output_conv.1.weight_g'] lowerCamelCase__ : Dict = checkpoint['output_conv.1.weight_v'] lowerCamelCase__ : int = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , ) -> List[Any]: if config_path is not None: lowerCamelCase__ : List[Any] = SpeechTaHifiGanConfig.from_pretrained(_UpperCAmelCase ) else: lowerCamelCase__ : Any = SpeechTaHifiGanConfig() lowerCamelCase__ : Tuple = SpeechTaHifiGan(_UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = torch.load(_UpperCAmelCase ) load_weights(orig_checkpoint['model']['generator'] , _UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : List[str] = np.load(_UpperCAmelCase ) lowerCamelCase__ : Tuple = stats[0].reshape(-1 ) lowerCamelCase__ : Optional[Any] = stats[1].reshape(-1 ) lowerCamelCase__ : int = torch.from_numpy(_UpperCAmelCase ).float() lowerCamelCase__ : Dict = torch.from_numpy(_UpperCAmelCase ).float() model.save_pretrained(_UpperCAmelCase ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(_UpperCAmelCase ) if __name__ == "__main__": _UpperCAmelCase : List[Any] = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _UpperCAmelCase : Optional[Any] = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
50
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging __snake_case : Dict = logging.get_logger(__name__) class A__(a_ ): """simple docstring""" _A : Dict = ['''pixel_values'''] def __init__( self , _lowercase = True , _lowercase = 1 / 255 , _lowercase = True , _lowercase = 8 , **_lowercase , ) -> None: super().__init__(**_lowercase ) a_ : Tuple = do_rescale a_ : Dict = rescale_factor a_ : int = do_pad a_ : Optional[int] = pad_size def UpperCamelCase__ ( self , _lowercase , _lowercase , _lowercase = None , **_lowercase ) -> np.ndarray: return rescale(_lowercase , scale=_lowercase , data_format=_lowercase , **_lowercase ) def UpperCamelCase__ ( self , _lowercase , _lowercase , _lowercase = None ) -> int: a_ , a_ : str = get_image_size(_lowercase ) a_ : Tuple = (old_height // size + 1) * size - old_height a_ : List[Any] = (old_width // size + 1) * size - old_width return pad(_lowercase , ((0, pad_height), (0, pad_width)) , mode="""symmetric""" , data_format=_lowercase ) def UpperCamelCase__ ( self , _lowercase , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = ChannelDimension.FIRST , **_lowercase , ) -> List[str]: a_ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale a_ : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor a_ : Tuple = do_pad if do_pad is not None else self.do_pad a_ : Tuple = pad_size if pad_size is not None else self.pad_size a_ : Tuple = make_list_of_images(_lowercase ) if not valid_images(_lowercase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) # All transformations expect numpy arrays. a_ : Tuple = [to_numpy_array(_lowercase ) for image in images] if do_rescale: a_ : Dict = [self.rescale(image=_lowercase , scale=_lowercase ) for image in images] if do_pad: a_ : str = [self.pad(_lowercase , size=_lowercase ) for image in images] a_ : Optional[int] = [to_channel_dimension_format(_lowercase , _lowercase ) for image in images] a_ : Optional[Any] = {"""pixel_values""": images} return BatchFeature(data=_lowercase , tensor_type=_lowercase )
248
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class a__ ( __snake_case ): A__ : Dict = 'philschmid/bart-large-cnn-samsum' A__ : Optional[Any] = ( 'This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, ' 'and returns a summary of the text.' ) A__ : int = 'summarizer' A__ : str = AutoTokenizer A__ : Dict = AutoModelForSeqaSeqLM A__ : List[str] = ['text'] A__ : Optional[Any] = ['text'] def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase ) -> List[str]: return self.pre_processor(UpperCAmelCase , return_tensors='pt' , truncation=UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase ) -> List[Any]: return self.model.generate(**UpperCAmelCase )[0] def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase ) -> Any: return self.pre_processor.decode(UpperCAmelCase , skip_special_tokens=UpperCAmelCase , clean_up_tokenization_spaces=UpperCAmelCase )
367
def lowerCAmelCase( __lowerCamelCase ): if not all(char in '01' for char in bin_string ): raise ValueError('Non-binary value was passed to the function' ) if not bin_string: raise ValueError('Empty string was passed to the function' ) __a = '' while len(__lowerCamelCase ) % 3 != 0: __a = '0' + bin_string __a = [ bin_string[index : index + 3] for index in range(len(__lowerCamelCase ) ) if index % 3 == 0 ] for bin_group in bin_string_in_3_list: __a = 0 for index, val in enumerate(__lowerCamelCase ): oct_val += int(2 ** (2 - index) * int(__lowerCamelCase ) ) oct_string += str(__lowerCamelCase ) return oct_string if __name__ == "__main__": from doctest import testmod testmod()
197
0
"""simple docstring""" def UpperCAmelCase__ ( lowerCAmelCase__ :int = 4_0_0_0_0_0_0 ) -> int: '''simple docstring''' lowercase = [] lowercase , lowercase = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(lowerCAmelCase__ ) lowercase , lowercase = b, a + b return sum(lowerCAmelCase__ ) if __name__ == "__main__": print(F"""{solution() = }""")
197
"""simple docstring""" from __future__ import annotations from collections.abc import Callable __lowerCAmelCase : str =list[list[float | int]] def UpperCAmelCase__ ( lowerCAmelCase__ :Matrix , lowerCAmelCase__ :Matrix ) -> Matrix: '''simple docstring''' lowercase = len(lowerCAmelCase__ ) lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowerCAmelCase__ )] lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 for row in range(lowerCAmelCase__ ): for col in range(lowerCAmelCase__ ): lowercase = matrix[row][col] lowercase = vector[row][0] lowercase = 0 lowercase = 0 while row < size and col < size: # pivoting lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowerCAmelCase__ , lowerCAmelCase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: lowercase , lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowerCAmelCase__ ): lowercase = augmented[rowa][col] / augmented[row][col] lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowerCAmelCase__ ): for row in range(lowerCAmelCase__ ): lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowerCAmelCase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 1_0 )] for row in range(lowerCAmelCase__ ) ] def UpperCAmelCase__ ( lowerCAmelCase__ :list[int] ) -> Callable[[int], int]: '''simple docstring''' lowercase = len(lowerCAmelCase__ ) lowercase = [[0 for _ in range(lowerCAmelCase__ )] for _ in range(lowerCAmelCase__ )] lowercase = [[0] for _ in range(lowerCAmelCase__ )] lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 for x_val, y_val in enumerate(lowerCAmelCase__ ): for col in range(lowerCAmelCase__ ): lowercase = (x_val + 1) ** (size - col - 1) lowercase = y_val lowercase = solve(lowerCAmelCase__ , lowerCAmelCase__ ) def interpolated_func(lowerCAmelCase__ :int ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowerCAmelCase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowerCAmelCase__ :int ) -> int: '''simple docstring''' return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**1_0 ) def UpperCAmelCase__ ( lowerCAmelCase__ :Callable[[int], int] = question_function , lowerCAmelCase__ :int = 1_0 ) -> int: '''simple docstring''' lowercase = [func(lowerCAmelCase__ ) for x_val in range(1 , order + 1 )] lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] lowercase = 0 lowercase = 42 lowercase = 42 for poly in polynomials: lowercase = 1 while func(lowerCAmelCase__ ) == poly(lowerCAmelCase__ ): x_val += 1 ret += poly(lowerCAmelCase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
197
1
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : Optional[int] = logging.get_logger(__name__) __UpperCamelCase : Dict = { '''huggingface/informer-tourism-monthly''': ( '''https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json''' ), # See all Informer models at https://huggingface.co/models?filter=informer } class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = "informer" lowercase__ = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", "num_hidden_layers": "encoder_layers", } def __init__( self : str ,lowercase_ : Optional[int] = None ,lowercase_ : Optional[int] = None ,lowercase_ : str = "student_t" ,lowercase_ : str = "nll" ,lowercase_ : int = 1 ,lowercase_ : List[int] = None ,lowercase_ : Optional[Union[str, bool]] = "mean" ,lowercase_ : int = 0 ,lowercase_ : int = 0 ,lowercase_ : int = 0 ,lowercase_ : int = 0 ,lowercase_ : Optional[List[int]] = None ,lowercase_ : Optional[List[int]] = None ,lowercase_ : int = 6_4 ,lowercase_ : int = 3_2 ,lowercase_ : int = 3_2 ,lowercase_ : int = 2 ,lowercase_ : int = 2 ,lowercase_ : int = 2 ,lowercase_ : int = 2 ,lowercase_ : bool = True ,lowercase_ : str = "gelu" ,lowercase_ : float = 0.05 ,lowercase_ : float = 0.1 ,lowercase_ : float = 0.1 ,lowercase_ : float = 0.1 ,lowercase_ : float = 0.1 ,lowercase_ : int = 1_0_0 ,lowercase_ : float = 0.02 ,lowercase_ : Union[str, Any]=True ,lowercase_ : str = "prob" ,lowercase_ : int = 5 ,lowercase_ : bool = True ,**lowercase_ : List[str] ,): # time series specific configuration lowerCAmelCase__ : Union[str, Any] = prediction_length lowerCAmelCase__ : str = context_length or prediction_length lowerCAmelCase__ : Any = distribution_output lowerCAmelCase__ : Optional[int] = loss lowerCAmelCase__ : Optional[int] = input_size lowerCAmelCase__ : str = num_time_features lowerCAmelCase__ : Dict = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ : Dict = scaling lowerCAmelCase__ : str = num_dynamic_real_features lowerCAmelCase__ : List[str] = num_static_real_features lowerCAmelCase__ : List[str] = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(lowercase_ ) != num_static_categorical_features: raise ValueError( '''The cardinality should be a list of the same length as `num_static_categorical_features`''' ) lowerCAmelCase__ : List[Any] = cardinality else: lowerCAmelCase__ : str = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(lowercase_ ) != num_static_categorical_features: raise ValueError( '''The embedding dimension should be a list of the same length as `num_static_categorical_features`''' ) lowerCAmelCase__ : List[str] = embedding_dimension else: lowerCAmelCase__ : Any = [min(5_0 ,(cat + 1) // 2 ) for cat in self.cardinality] lowerCAmelCase__ : List[Any] = num_parallel_samples # Transformer architecture configuration lowerCAmelCase__ : int = input_size * len(self.lags_sequence ) + self._number_of_features lowerCAmelCase__ : List[Any] = d_model lowerCAmelCase__ : int = encoder_attention_heads lowerCAmelCase__ : Union[str, Any] = decoder_attention_heads lowerCAmelCase__ : int = encoder_ffn_dim lowerCAmelCase__ : Tuple = decoder_ffn_dim lowerCAmelCase__ : Optional[int] = encoder_layers lowerCAmelCase__ : Optional[int] = decoder_layers lowerCAmelCase__ : Any = dropout lowerCAmelCase__ : List[Any] = attention_dropout lowerCAmelCase__ : List[Any] = activation_dropout lowerCAmelCase__ : List[Any] = encoder_layerdrop lowerCAmelCase__ : List[str] = decoder_layerdrop lowerCAmelCase__ : Union[str, Any] = activation_function lowerCAmelCase__ : int = init_std lowerCAmelCase__ : List[Any] = use_cache # Informer lowerCAmelCase__ : Any = attention_type lowerCAmelCase__ : int = sampling_factor lowerCAmelCase__ : Dict = distil super().__init__(is_encoder_decoder=lowercase_ ,**lowercase_ ) @property def __lowerCAmelCase ( self : Optional[int] ): return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
354
"""simple docstring""" import math def __SCREAMING_SNAKE_CASE ( A_ , A_ = 0 , A_ = 0 ): lowerCAmelCase__ : Tuple = end or len(A_ ) for i in range(A_ , A_ ): lowerCAmelCase__ : Dict = i lowerCAmelCase__ : List[Any] = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: lowerCAmelCase__ : str = array[temp_index - 1] temp_index -= 1 lowerCAmelCase__ : int = temp_index_value return array def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ ): # Max Heap lowerCAmelCase__ : List[str] = index lowerCAmelCase__ : Any = 2 * index + 1 # Left Node lowerCAmelCase__ : int = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: lowerCAmelCase__ : Tuple = left_index if right_index < heap_size and array[largest] < array[right_index]: lowerCAmelCase__ : int = right_index if largest != index: lowerCAmelCase__ ,lowerCAmelCase__ : Dict = array[largest], array[index] heapify(A_ , A_ , A_ ) def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Any = len(A_ ) for i in range(n // 2 , -1 , -1 ): heapify(A_ , A_ , A_ ) for i in range(n - 1 , 0 , -1 ): lowerCAmelCase__ ,lowerCAmelCase__ : List[str] = array[0], array[i] heapify(A_ , 0 , A_ ) return array def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ ): if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ ): lowerCAmelCase__ : str = low lowerCAmelCase__ : Union[str, Any] = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i lowerCAmelCase__ ,lowerCAmelCase__ : Union[str, Any] = array[j], array[i] i += 1 def __SCREAMING_SNAKE_CASE ( A_ ): if len(A_ ) == 0: return array lowerCAmelCase__ : int = 2 * math.ceil(math.loga(len(A_ ) ) ) lowerCAmelCase__ : Optional[Any] = 16 return intro_sort(A_ , 0 , len(A_ ) , A_ , A_ ) def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ , A_ ): while end - start > size_threshold: if max_depth == 0: return heap_sort(A_ ) max_depth -= 1 lowerCAmelCase__ : List[str] = median_of_a(A_ , A_ , start + ((end - start) // 2) + 1 , end - 1 ) lowerCAmelCase__ : Union[str, Any] = partition(A_ , A_ , A_ , A_ ) intro_sort(A_ , A_ , A_ , A_ , A_ ) lowerCAmelCase__ : Optional[int] = p return insertion_sort(A_ , A_ , A_ ) if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase : Optional[Any] = input('''Enter numbers separated by a comma : ''').strip() __UpperCamelCase : Tuple = [float(item) for item in user_input.split(''',''')] print(sort(unsorted))
74
0
from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 __A = { # 1536-bit 5: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 2048-bit 14: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AACAA68FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 3072-bit 15: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 4096-bit 16: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7" + "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA" + "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6" + "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED" + "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9" + "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199" + "FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 6144-bit 17: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08" + "8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B" + "302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9" + "A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6" + "49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8" + "FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C" + "180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718" + "3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D" + "04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D" + "B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226" + "1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC" + "E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26" + "99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB" + "04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2" + "233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127" + "D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492" + "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406" + "AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918" + "DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151" + "2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03" + "F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F" + "BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA" + "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B" + "B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632" + "387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E" + "6DCC4024FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 8192-bit 18: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7" + "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA" + "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6" + "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED" + "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9" + "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492" + "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD" + "F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831" + "179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B" + "DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF" + "5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6" + "D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3" + "23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA" + "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328" + "06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C" + "DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE" + "12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4" + "38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300" + "741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568" + "3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9" + "22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B" + "4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A" + "062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36" + "4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1" + "B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92" + "4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47" + "9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71" + "60C980DD98EDD3DFFFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, } class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self : int , UpperCAmelCase_ : int = 14) ->None: '''simple docstring''' if group not in primes: raise ValueError("Unsupported Group") lowerCamelCase__: Union[str, Any] =primes[group]["prime"] lowerCamelCase__: Optional[int] =primes[group]["generator"] lowerCamelCase__: Dict =int(hexlify(urandom(32)) , base=16) def SCREAMING_SNAKE_CASE_ (self : str) ->str: '''simple docstring''' return hex(self.__private_key)[2:] def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->str: '''simple docstring''' lowerCamelCase__: str =pow(self.generator , self.__private_key , self.prime) return hex(UpperCAmelCase_)[2:] def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : int) ->bool: '''simple docstring''' return ( 2 <= key <= self.prime - 2 and pow(UpperCAmelCase_ , (self.prime - 1) // 2 , self.prime) == 1 ) def SCREAMING_SNAKE_CASE_ (self : Optional[int] , UpperCAmelCase_ : str) ->str: '''simple docstring''' lowerCamelCase__: int =int(UpperCAmelCase_ , base=16) if not self.is_valid_public_key(UpperCAmelCase_): raise ValueError("Invalid public key") lowerCamelCase__: Dict =pow(UpperCAmelCase_ , self.__private_key , self.prime) return shaaaa(str(UpperCAmelCase_).encode()).hexdigest() @staticmethod def SCREAMING_SNAKE_CASE_ (UpperCAmelCase_ : int , UpperCAmelCase_ : int) ->bool: '''simple docstring''' return ( 2 <= remote_public_key_str <= prime - 2 and pow(UpperCAmelCase_ , (prime - 1) // 2 , UpperCAmelCase_) == 1 ) @staticmethod def SCREAMING_SNAKE_CASE_ (UpperCAmelCase_ : str , UpperCAmelCase_ : str , UpperCAmelCase_ : int = 14) ->str: '''simple docstring''' lowerCamelCase__: Dict =int(UpperCAmelCase_ , base=16) lowerCamelCase__: str =int(UpperCAmelCase_ , base=16) lowerCamelCase__: Union[str, Any] =primes[group]["prime"] if not DiffieHellman.is_valid_public_key_static(UpperCAmelCase_ , UpperCAmelCase_): raise ValueError("Invalid public key") lowerCamelCase__: List[Any] =pow(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_) return shaaaa(str(UpperCAmelCase_).encode()).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
10
'''simple docstring''' import copy import re class A__ : A__ = 'hp' A__ = {} A__ = None @classmethod def A ( cls : Optional[Any] , _a : Optional[Any] , _a : Any ) -> Union[str, Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =prefix _SCREAMING_SNAKE_CASE =defaults cls.build_naming_info() @staticmethod def A ( _a : Optional[Any] , _a : List[Any] ) -> Any: '''simple docstring''' if len(_a ) == 0: return "" _SCREAMING_SNAKE_CASE =None if any(char.isdigit() for char in word ): raise Exception(f"Parameters should not contain numbers: '{word}' contains a number" ) if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(_a ) + 1 ): _SCREAMING_SNAKE_CASE =word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: _SCREAMING_SNAKE_CASE =prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(_a : str ): _SCREAMING_SNAKE_CASE ='' while integer != 0: _SCREAMING_SNAKE_CASE =chr(ord('A' ) + integer % 10 ) + s integer //= 10 return s _SCREAMING_SNAKE_CASE =0 while True: _SCREAMING_SNAKE_CASE =word + '#' + int_to_alphabetic(_a ) if sword in info["reverse_short_word"]: continue else: _SCREAMING_SNAKE_CASE =sword break _SCREAMING_SNAKE_CASE =short_word _SCREAMING_SNAKE_CASE =word return short_word @staticmethod def A ( _a : Optional[Any] , _a : int ) -> Optional[int]: '''simple docstring''' _SCREAMING_SNAKE_CASE =param_name.split('_' ) _SCREAMING_SNAKE_CASE =[TrialShortNamer.shortname_for_word(_a , _a ) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name _SCREAMING_SNAKE_CASE =['', '_'] for separator in separators: _SCREAMING_SNAKE_CASE =separator.join(_a ) if shortname not in info["reverse_short_param"]: _SCREAMING_SNAKE_CASE =shortname _SCREAMING_SNAKE_CASE =param_name return shortname return param_name @staticmethod def A ( _a : Dict , _a : int ) -> Optional[Any]: '''simple docstring''' _SCREAMING_SNAKE_CASE =TrialShortNamer.shortname_for_key(_a , _a ) _SCREAMING_SNAKE_CASE =short_name _SCREAMING_SNAKE_CASE =param_name @classmethod def A ( cls : Optional[int] ) -> Tuple: '''simple docstring''' if cls.NAMING_INFO is not None: return _SCREAMING_SNAKE_CASE ={ 'short_word': {}, 'reverse_short_word': {}, 'short_param': {}, 'reverse_short_param': {}, } _SCREAMING_SNAKE_CASE =list(cls.DEFAULTS.keys() ) for k in field_keys: cls.add_new_param_name(_a , _a ) _SCREAMING_SNAKE_CASE =info @classmethod def A ( cls : List[Any] , _a : int ) -> int: '''simple docstring''' cls.build_naming_info() assert cls.PREFIX is not None _SCREAMING_SNAKE_CASE =[copy.copy(cls.PREFIX )] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(f"You should provide a default value for the param name {k} with value {v}" ) if v == cls.DEFAULTS[k]: # The default value is not added to the name continue _SCREAMING_SNAKE_CASE =cls.NAMING_INFO['short_param'][k] if isinstance(_a , _a ): _SCREAMING_SNAKE_CASE =1 if v else 0 _SCREAMING_SNAKE_CASE ='' if isinstance(_a , (int, float) ) else '-' _SCREAMING_SNAKE_CASE =f"{key}{sep}{v}" name.append(_a ) return "_".join(_a ) @classmethod def A ( cls : Optional[Any] , _a : List[Any] ) -> Dict: '''simple docstring''' _SCREAMING_SNAKE_CASE =repr[len(cls.PREFIX ) + 1 :] if repr == "": _SCREAMING_SNAKE_CASE =[] else: _SCREAMING_SNAKE_CASE =repr.split('_' ) _SCREAMING_SNAKE_CASE ={} for value in values: if "-" in value: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =value.split('-' ) else: _SCREAMING_SNAKE_CASE =re.sub('[0-9.]' , '' , _a ) _SCREAMING_SNAKE_CASE =float(re.sub('[^0-9.]' , '' , _a ) ) _SCREAMING_SNAKE_CASE =cls.NAMING_INFO['reverse_short_param'][p_k] _SCREAMING_SNAKE_CASE =p_v for k in cls.DEFAULTS: if k not in parameters: _SCREAMING_SNAKE_CASE =cls.DEFAULTS[k] return parameters
47
0
"""simple docstring""" import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __lowerCamelCase : @staticmethod def A__ (*lowerCamelCase , **lowerCamelCase ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_torch class __lowerCamelCase ( unittest.TestCase ): __UpperCamelCase = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def A__ (self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): '''simple docstring''' _lowerCAmelCase = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) _lowerCAmelCase = [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] return object_detector, examples def A__ (self , lowerCamelCase , lowerCamelCase ): '''simple docstring''' _lowerCAmelCase = object_detector(examples[0] , threshold=0.0 ) _lowerCAmelCase = len(lowerCamelCase ) self.assertGreater(lowerCamelCase , 0 ) self.assertEqual( lowerCamelCase , [ { """score""": ANY(lowerCamelCase ), """label""": ANY(lowerCamelCase ), """box""": {"""xmin""": ANY(lowerCamelCase ), """ymin""": ANY(lowerCamelCase ), """xmax""": ANY(lowerCamelCase ), """ymax""": ANY(lowerCamelCase )}, } for i in range(lowerCamelCase ) ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def A__ (self ): '''simple docstring''' pass @require_torch def A__ (self ): '''simple docstring''' _lowerCAmelCase = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) _lowerCAmelCase = object_detector( """./tests/fixtures/tests_samples/COCO/000000039769.png""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=0.64 , ) self.assertEqual( nested_simplify(lowerCamelCase , decimals=4 ) , [ {"""score""": 0.7235, """label""": """cat""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7218, """label""": """remote""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7184, """label""": """couch""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.6748, """label""": """remote""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6656, """label""": """cat""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6614, """label""": """couch""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6456, """label""": """remote""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, {"""score""": 0.642, """label""": """remote""", """box""": {"""xmin""": 67, """ymin""": 274, """xmax""": 93, """ymax""": 297}}, {"""score""": 0.6419, """label""": """cat""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, ] , ) _lowerCAmelCase = object_detector( [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(lowerCamelCase , decimals=4 ) , [ [ {"""score""": 0.7235, """label""": """cat""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7218, """label""": """remote""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.7184, """label""": """couch""", """box""": {"""xmin""": 204, """ymin""": 167, """xmax""": 232, """ymax""": 190}}, {"""score""": 0.6748, """label""": """remote""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6656, """label""": """cat""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6614, """label""": """couch""", """box""": {"""xmin""": 571, """ymin""": 83, """xmax""": 598, """ymax""": 103}}, {"""score""": 0.6456, """label""": """remote""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, {"""score""": 0.642, """label""": """remote""", """box""": {"""xmin""": 67, """ymin""": 274, """xmax""": 93, """ymax""": 297}}, {"""score""": 0.6419, """label""": """cat""", """box""": {"""xmin""": 494, """ymin""": 105, """xmax""": 521, """ymax""": 127}}, ] ] , ) @require_torch @slow def A__ (self ): '''simple docstring''' _lowerCAmelCase = pipeline("""zero-shot-object-detection""" ) _lowerCAmelCase = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , ) self.assertEqual( nested_simplify(lowerCamelCase , decimals=4 ) , [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, {"""score""": 0.1474, """label""": """remote""", """box""": {"""xmin""": 335, """ymin""": 74, """xmax""": 371, """ymax""": 187}}, {"""score""": 0.1208, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 642, """ymax""": 476}}, ] , ) _lowerCAmelCase = object_detector( [ { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, ] , ) self.assertEqual( nested_simplify(lowerCamelCase , decimals=4 ) , [ [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, {"""score""": 0.1474, """label""": """remote""", """box""": {"""xmin""": 335, """ymin""": 74, """xmax""": 371, """ymax""": 187}}, {"""score""": 0.1208, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 642, """ymax""": 476}}, ], [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, {"""score""": 0.1474, """label""": """remote""", """box""": {"""xmin""": 335, """ymin""": 74, """xmax""": 371, """ymax""": 187}}, {"""score""": 0.1208, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 642, """ymax""": 476}}, ], ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def A__ (self ): '''simple docstring''' pass @require_torch @slow def A__ (self ): '''simple docstring''' _lowerCAmelCase = 0.2 _lowerCAmelCase = pipeline("""zero-shot-object-detection""" ) _lowerCAmelCase = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=lowerCamelCase , ) self.assertEqual( nested_simplify(lowerCamelCase , decimals=4 ) , [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, {"""score""": 0.2537, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 55, """xmax""": 315, """ymax""": 472}}, ] , ) @require_torch @slow def A__ (self ): '''simple docstring''' _lowerCAmelCase = 2 _lowerCAmelCase = pipeline("""zero-shot-object-detection""" ) _lowerCAmelCase = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , top_k=lowerCamelCase , ) self.assertEqual( nested_simplify(lowerCamelCase , decimals=4 ) , [ {"""score""": 0.2868, """label""": """cat""", """box""": {"""xmin""": 324, """ymin""": 20, """xmax""": 640, """ymax""": 373}}, {"""score""": 0.277, """label""": """remote""", """box""": {"""xmin""": 40, """ymin""": 72, """xmax""": 177, """ymax""": 115}}, ] , )
317
"""simple docstring""" def __UpperCAmelCase ( snake_case_ : int = 1000000 ) -> int: """simple docstring""" _lowerCAmelCase = limit + 1 _lowerCAmelCase = [0] * limit for first_term in range(1 , snake_case_ ): for n in range(snake_case_ , snake_case_ , snake_case_ ): _lowerCAmelCase = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _lowerCAmelCase = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'{solution() = }')
317
1
'''simple docstring''' import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging _lowerCamelCase : Dict = logging.get_logger(__name__) def __a ( UpperCAmelCase ) ->List[str]: """simple docstring""" A = r'\w+[.]\d+' A = re.findall(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for pat in pats: A = key.replace(_SCREAMING_SNAKE_CASE , """_""".join(pat.split(""".""" ) ) ) return key def __a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->str: """simple docstring""" A = pt_tuple_key[:-1] + ('scale',) if ( any("""norm""" in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): A = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: A = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: A = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer A = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: A = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer A = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": A = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight A = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias A = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def __a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=42 ) ->List[str]: """simple docstring""" A = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params A = flax_model.init_weights(PRNGKey(_SCREAMING_SNAKE_CASE ) ) A = flatten_dict(_SCREAMING_SNAKE_CASE ) A = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): A = rename_key(_SCREAMING_SNAKE_CASE ) A = tuple(renamed_pt_key.split(""".""" ) ) # Correctly rename weight parameters A = rename_key_and_reshape_tensor(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown A = jnp.asarray(_SCREAMING_SNAKE_CASE ) return unflatten_dict(_SCREAMING_SNAKE_CASE )
258
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: Optional[int] = SamImageProcessor() a__: Tuple = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> List[Any]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[Any] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Optional[int] = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: List[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = self.get_image_processor() a__: List[Any] = SamProcessor(image_processor=lowercase) a__: Optional[int] = self.prepare_image_inputs() a__: Optional[Any] = image_processor(lowercase , return_tensors='np') a__: Tuple = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_torch def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = self.get_image_processor() a__: List[str] = SamProcessor(image_processor=lowercase) a__: Optional[Any] = [torch.ones((1, 3, 5, 5))] a__: Union[str, Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: int = processor.post_process_masks(lowercase , lowercase , lowercase) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Optional[int] = processor.post_process_masks( lowercase , torch.tensor(lowercase) , torch.tensor(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Dict = [np.ones((1, 3, 5, 5))] a__: Tuple = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = [[1, 0], [0, 1]] with self.assertRaises(lowercase): a__: List[Any] = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) @require_vision @require_tf class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: List[Any] = SamImageProcessor() a__: Optional[int] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> int: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[int] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Dict = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = self.get_image_processor() a__: str = SamProcessor(image_processor=lowercase) a__: int = self.prepare_image_inputs() a__: int = image_processor(lowercase , return_tensors='np') a__: Dict = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_tf def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Any = SamProcessor(image_processor=lowercase) a__: str = [tf.ones((1, 3, 5, 5))] a__: List[Any] = [[17_64, 26_46]] a__: List[Any] = [[6_83, 10_24]] a__: List[Any] = processor.post_process_masks(lowercase , lowercase , lowercase , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = processor.post_process_masks( lowercase , tf.convert_to_tensor(lowercase) , tf.convert_to_tensor(lowercase) , return_tensors='tf' , ) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Optional[Any] = [np.ones((1, 3, 5, 5))] a__: int = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: List[str] = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError): a__: Any = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = tempfile.mkdtemp() a__: int = SamImageProcessor() a__: Union[str, Any] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Any = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = self.get_image_processor() a__: int = SamProcessor(image_processor=lowercase) a__: int = np.random.randint(0 , 2 , size=(1, 3, 5, 5)).astype(np.floataa) a__: Dict = [tf.convert_to_tensor(lowercase)] a__: Union[str, Any] = [torch.tensor(lowercase)] a__: List[Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: Tuple = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='tf') a__: str = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='pt') self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy())) @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Dict = SamProcessor(image_processor=lowercase) a__: Any = self.prepare_image_inputs() a__: List[Any] = image_processor(lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Tuple = processor(images=lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Any = image_processor(lowercase , return_tensors='tf')['pixel_values'].numpy() a__: Any = processor(images=lowercase , return_tensors='tf')['pixel_values'].numpy() self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase))
290
0
"""simple docstring""" def _snake_case ( lowerCamelCase__ : int , lowerCamelCase__ : int ) -> Union[str, Any]: if b == 0: return 1 if (b % 2) == 0: return actual_power(lowerCamelCase__ , int(b / 2 ) ) * actual_power(lowerCamelCase__ , int(b / 2 ) ) else: return a * actual_power(lowerCamelCase__ , int(b / 2 ) ) * actual_power(lowerCamelCase__ , int(b / 2 ) ) def _snake_case ( lowerCamelCase__ : int , lowerCamelCase__ : int ) -> float: if b < 0: return 1 / actual_power(lowerCamelCase__ , lowerCamelCase__ ) return actual_power(lowerCamelCase__ , lowerCamelCase__ ) if __name__ == "__main__": print(power(-2, -3))
209
"""simple docstring""" import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline A__ : int = { 'n_samples': 64, 'horizon': 32, 'num_inference_steps': 20, 'n_guide_steps': 2, # can set to 0 for faster sampling, does not use value network 'scale_grad_by_std': True, 'scale': 0.1, 'eta': 0.0, 't_grad_cutoff': 2, 'device': 'cpu', } if __name__ == "__main__": A__ : str = 'hopper-medium-v2' A__ : Dict = gym.make(env_name) A__ : List[Any] = ValueGuidedRLPipeline.from_pretrained( 'bglick13/hopper-medium-v2-value-function-hor32', env=env, ) env.seed(0) A__ : Dict = env.reset() A__ : Optional[int] = 0 A__ : str = 0 A__ : List[Any] = 1_000 A__ : Tuple = [obs.copy()] try: for t in tqdm.tqdm(range(T)): # call the policy A__ : Union[str, Any] = pipeline(obs, planning_horizon=32) # execute action in environment A__ , A__ , A__ , A__ : Any = env.step(denorm_actions) A__ : List[Any] = env.get_normalized_score(total_reward) # update return total_reward += reward total_score += score print( f'Step: {t}, Reward: {reward}, Total Reward: {total_reward}, Score: {score}, Total Score:' f' {total_score}' ) # save observations for rendering rollout.append(next_observation.copy()) A__ : Optional[Any] = next_observation except KeyboardInterrupt: pass print(f'Total reward: {total_reward}')
209
1
from collections.abc import Sequence def lowerCAmelCase_ ( __lowerCAmelCase = None )-> int: '''simple docstring''' if nums is None or not nums: raise ValueError('''Input sequence should not be empty''' ) UpperCAmelCase : Dict =nums[0] for i in range(1 , len(__lowerCAmelCase ) ): UpperCAmelCase : Dict =nums[i] UpperCAmelCase : Any =max(__lowerCAmelCase , ans + num , __lowerCAmelCase ) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user __snake_case = int(input('''Enter number of elements : ''').strip()) __snake_case = list(map(int, input('''\nEnter the numbers : ''').strip().split()))[:n] print(max_subsequence_sum(array))
348
from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowerCAmelCase_ ( )-> int: '''simple docstring''' UpperCAmelCase : str ={ '''repo_name''': ['''test_repo1''', '''test_repo2''', '''test_repo3'''], '''path''': ['''test_1.py''', '''test_2.py''', '''unit_test.py'''], '''content''': ['''a ''' * 20, '''a ''' * 30, '''b ''' * 7], } UpperCAmelCase : Union[str, Any] =Dataset.from_dict(__lowerCAmelCase ) return dataset class __snake_case ( lowerCamelCase__ ): def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : List[str] =get_dataset() UpperCAmelCase : Optional[int] =make_duplicate_clusters(snake_case__ , 0.85 ) self.assertEqual(len(duplicate_clusters[0] ) , 2 ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : str =get_dataset() UpperCAmelCase , UpperCAmelCase : Tuple =deduplicate_dataset(snake_case__ ) self.assertEqual(len(snake_case__ ) , 2 ) print(snake_case__ ) self.assertEqual(duplicate_clusters[0][0]['''copies'''] , 2 ) self.assertEqual(duplicate_clusters[0][0]['''is_extreme'''] , snake_case__ )
348
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCamelCase__ = logging.get_logger(__name__) class __magic_name__ (__lowercase ): lowerCamelCase__ = ['''pixel_values'''] def __init__( self , _a = True , _a = None , _a = None , _a = PILImageResampling.BILINEAR , _a = True , _a = 1 / 255 , _a = True , _a = None , _a = None , **_a , ) -> None: super().__init__(**_a ) lowerCAmelCase_ = size if size is not None else {"shortest_edge": 384} lowerCAmelCase_ = get_size_dict(_a , default_to_square=_a ) lowerCAmelCase_ = do_resize lowerCAmelCase_ = size # Default value set here for backwards compatibility where the value in config is None lowerCAmelCase_ = crop_pct if crop_pct is not None else 224 / 256 lowerCAmelCase_ = resample lowerCAmelCase_ = do_rescale lowerCAmelCase_ = rescale_factor lowerCAmelCase_ = do_normalize lowerCAmelCase_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCAmelCase_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def __a ( self , _a , _a , _a , _a = PILImageResampling.BICUBIC , _a = None , **_a , ) -> np.ndarray: lowerCAmelCase_ = get_size_dict(_a , default_to_square=_a ) if "shortest_edge" not in size: raise ValueError(f"Size dictionary must contain 'shortest_edge' key. Got {size.keys()}" ) lowerCAmelCase_ = size["shortest_edge"] if shortest_edge < 384: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCAmelCase_ = int(shortest_edge / crop_pct ) lowerCAmelCase_ = get_resize_output_image_size(_a , size=_a , default_to_square=_a ) lowerCAmelCase_ = resize(image=_a , size=_a , resample=_a , data_format=_a , **_a ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=_a , size=(shortest_edge, shortest_edge) , data_format=_a , **_a ) else: # warping (no cropping) when evaluated at 384 or larger return resize( _a , size=(shortest_edge, shortest_edge) , resample=_a , data_format=_a , **_a ) def __a ( self , _a , _a , _a = None , **_a , ) -> Any: return rescale(_a , scale=_a , data_format=_a , **_a ) def __a ( self , _a , _a , _a , _a = None , **_a , ) -> np.ndarray: return normalize(_a , mean=_a , std=_a , data_format=_a , **_a ) def __a ( self , _a , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = ChannelDimension.FIRST , **_a , ) -> PIL.Image.Image: lowerCAmelCase_ = do_resize if do_resize is not None else self.do_resize lowerCAmelCase_ = crop_pct if crop_pct is not None else self.crop_pct lowerCAmelCase_ = resample if resample is not None else self.resample lowerCAmelCase_ = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase_ = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase_ = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase_ = image_mean if image_mean is not None else self.image_mean lowerCAmelCase_ = image_std if image_std is not None else self.image_std lowerCAmelCase_ = size if size is not None else self.size lowerCAmelCase_ = get_size_dict(_a , default_to_square=_a ) lowerCAmelCase_ = make_list_of_images(_a ) if not valid_images(_a ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_resize and size["shortest_edge"] < 384 and crop_pct is None: raise ValueError("crop_pct must be specified if size < 384." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. lowerCAmelCase_ = [to_numpy_array(_a ) for image in images] if do_resize: lowerCAmelCase_ = [self.resize(image=_a , size=_a , crop_pct=_a , resample=_a ) for image in images] if do_rescale: lowerCAmelCase_ = [self.rescale(image=_a , scale=_a ) for image in images] if do_normalize: lowerCAmelCase_ = [self.normalize(image=_a , mean=_a , std=_a ) for image in images] lowerCAmelCase_ = [to_channel_dimension_format(_a , _a ) for image in images] lowerCAmelCase_ = {"pixel_values": images} return BatchFeature(data=_a , tensor_type=_a )
22
def A(__a: Optional[Any] ): lowerCAmelCase_ = len(__a ) lowerCAmelCase_ = sum(__a ) lowerCAmelCase_ = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): lowerCAmelCase_ = True for i in range(1 , s + 1 ): lowerCAmelCase_ = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): lowerCAmelCase_ = dp[i][j - 1] if arr[i - 1] <= j: lowerCAmelCase_ = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: lowerCAmelCase_ = s - 2 * j break return diff
22
1
from __future__ import annotations def lowerCamelCase__ ( snake_case_ : list[int] , snake_case_ : int ) -> bool: if len(snake_case_ ) == 0: return False __snake_case = len(snake_case_ ) // 2 if a_list[midpoint] == item: return True if item < a_list[midpoint]: return binary_search(a_list[:midpoint] , snake_case_ ) else: return binary_search(a_list[midpoint + 1 :] , snake_case_ ) if __name__ == "__main__": snake_case_ = input('Enter numbers separated by comma:\n').strip() snake_case_ = [int(item.strip()) for item in user_input.split(',')] snake_case_ = int(input('Enter the number to be found in the list:\n').strip()) snake_case_ = '' if binary_search(sequence, target) else 'not ' print(F'{target} was {not_str}found in {sequence}')
24
import re import string import numpy as np import datasets _SCREAMING_SNAKE_CASE : Union[str, Any] = "\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n" _SCREAMING_SNAKE_CASE : Tuple = "\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results[\"exact_match\"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"]\n >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results[\"exact_match\"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric(\"exact_match\")\n >>> refs = [\"The cat sat on the mat.\", \"Theaters are great.\", \"It's like comparing oranges and apples.\"]\n >>> preds = [\"The cat sat on the mat?\", \"Theaters are great.\", \"It's like comparing apples and oranges.\"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results[\"exact_match\"], 1))\n 33.3\n\n" _SCREAMING_SNAKE_CASE : Any = "\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): """simple docstring""" def a_ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , reference_urls=[] , ) def a_ ( self , __snake_case , __snake_case , __snake_case=None , __snake_case=False , __snake_case=False , __snake_case=False , ): if regexes_to_ignore is not None: for s in regexes_to_ignore: snake_case = np.array([re.sub(__snake_case , '''''' , __snake_case ) for x in predictions] ) snake_case = np.array([re.sub(__snake_case , '''''' , __snake_case ) for x in references] ) else: snake_case = np.asarray(__snake_case ) snake_case = np.asarray(__snake_case ) if ignore_case: snake_case = np.char.lower(__snake_case ) snake_case = np.char.lower(__snake_case ) if ignore_punctuation: snake_case = string.punctuation.maketrans('''''' , '''''' , string.punctuation ) snake_case = np.char.translate(__snake_case , table=__snake_case ) snake_case = np.char.translate(__snake_case , table=__snake_case ) if ignore_numbers: snake_case = string.digits.maketrans('''''' , '''''' , string.digits ) snake_case = np.char.translate(__snake_case , table=__snake_case ) snake_case = np.char.translate(__snake_case , table=__snake_case ) snake_case = predictions == references return {"exact_match": np.mean(__snake_case ) * 1_0_0}
127
0
import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCAmelCase__ ( A__ , unittest.TestCase ): """simple docstring""" a = VideoToVideoSDPipeline a = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({"video"} ) - {"image", "width", "height"} a = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"video"} ) - {"image"} a = PipelineTesterMixin.required_optional_params - {"latents"} a = False # No `output_type`. a = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def lowercase_ ( self : str ) -> Tuple: torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) SCREAMING_SNAKE_CASE__ = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=__lowerCamelCase , set_alpha_to_one=__lowerCamelCase , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=512 , ) SCREAMING_SNAKE_CASE__ = CLIPTextModel(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) SCREAMING_SNAKE_CASE__ = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def lowercase_ ( self : str , __lowerCamelCase : int , __lowerCamelCase : Tuple=0 ) -> List[str]: # 3 frames SCREAMING_SNAKE_CASE__ = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(__lowerCamelCase ) ).to(__lowerCamelCase ) if str(__lowerCamelCase ).startswith('''mps''' ): SCREAMING_SNAKE_CASE__ = torch.manual_seed(__lowerCamelCase ) else: SCREAMING_SNAKE_CASE__ = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''video''': video, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def lowercase_ ( self : Optional[int] ) -> str: SCREAMING_SNAKE_CASE__ = '''cpu''' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = VideoToVideoSDPipeline(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = sd_pipe.to(__lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = '''np''' SCREAMING_SNAKE_CASE__ = sd_pipe(**__lowerCamelCase ).frames SCREAMING_SNAKE_CASE__ = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) SCREAMING_SNAKE_CASE__ = np.array([106, 117, 113, 174, 137, 112, 148, 151, 131] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def lowercase_ ( self : Tuple ) -> Tuple: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowerCamelCase , expected_max_diff=5e-3 ) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def lowercase_ ( self : Optional[int] ) -> List[str]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def lowercase_ ( self : List[str] ) -> Optional[int]: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''' ) def lowercase_ ( self : str ) -> Dict: pass def lowercase_ ( self : List[Any] ) -> int: return super().test_progress_bar() @slow @skip_mps class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self : Dict ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = VideoToVideoSDPipeline.from_pretrained('''cerspense/zeroscope_v2_XL''' , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames SCREAMING_SNAKE_CASE__ = torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ = torch.randn((1, 10, 3, 1024, 576) , generator=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = video.to('''cuda''' ) SCREAMING_SNAKE_CASE__ = '''Spiderman is surfing''' SCREAMING_SNAKE_CASE__ = pipe(__lowerCamelCase , video=__lowerCamelCase , generator=__lowerCamelCase , num_inference_steps=3 , output_type='''pt''' ).frames SCREAMING_SNAKE_CASE__ = np.array([-1.0458984, -1.1279297, -0.9663086, -0.91503906, -0.75097656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
360
import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class UpperCAmelCase__ ( A__ ): """simple docstring""" def __init__( self : List[str] , __lowerCamelCase : Tuple=0.01 , __lowerCamelCase : Optional[Any]=1000 ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = p_stop SCREAMING_SNAKE_CASE__ = max_length def __iter__( self : List[str] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = False while not stop and count < self.max_length: yield count count += 1 SCREAMING_SNAKE_CASE__ = random.random() < self.p_stop class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=True ) -> Dict: SCREAMING_SNAKE_CASE__ = [ BatchSamplerShard(__lowerCamelCase , 2 , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) for i in range(2 ) ] SCREAMING_SNAKE_CASE__ = [list(__lowerCamelCase ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(__lowerCamelCase ) for shard in batch_sampler_shards] , [len(__lowerCamelCase ) for e in expected] ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def lowercase_ ( self : Any ) -> Optional[Any]: # Check the shards when the dataset is a round multiple of total batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. SCREAMING_SNAKE_CASE__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is very small. SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) def lowercase_ ( self : Optional[int] ) -> int: # Check the shards when the dataset is a round multiple of batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) # Check the shards when the dataset is very small. SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) def lowercase_ ( self : str ) -> Dict: # Check the shards when the dataset is a round multiple of total batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. SCREAMING_SNAKE_CASE__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is very small. SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[[0, 1]], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) def lowercase_ ( self : List[str] ) -> Tuple: # Check the shards when the dataset is a round multiple of batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is very small. SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[[0, 1]], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) def lowercase_ ( self : Optional[int] ) -> List[str]: SCREAMING_SNAKE_CASE__ = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] SCREAMING_SNAKE_CASE__ = [BatchSamplerShard(__lowerCamelCase , 2 , __lowerCamelCase , even_batches=__lowerCamelCase ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def lowercase_ ( self : Optional[Any] , __lowerCamelCase : Any , __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : Tuple=False , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Dict=False ) -> str: random.seed(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = list(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = [ IterableDatasetShard( __lowerCamelCase , batch_size=__lowerCamelCase , drop_last=__lowerCamelCase , num_processes=__lowerCamelCase , process_index=__lowerCamelCase , split_batches=__lowerCamelCase , ) for i in range(__lowerCamelCase ) ] SCREAMING_SNAKE_CASE__ = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(__lowerCamelCase ) iterable_dataset_lists.append(list(__lowerCamelCase ) ) SCREAMING_SNAKE_CASE__ = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size SCREAMING_SNAKE_CASE__ = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(__lowerCamelCase ) , len(__lowerCamelCase ) ) self.assertTrue(len(__lowerCamelCase ) % shard_batch_size == 0 ) SCREAMING_SNAKE_CASE__ = [] for idx in range(0 , len(__lowerCamelCase ) , __lowerCamelCase ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(__lowerCamelCase ) < len(__lowerCamelCase ): reference += reference self.assertListEqual(__lowerCamelCase , reference[: len(__lowerCamelCase )] ) def lowercase_ ( self : str ) -> Dict: SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = RandomIterableDataset() self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) # Edge case with a very small dataset SCREAMING_SNAKE_CASE__ = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) def lowercase_ ( self : Optional[Any] ) -> Any: SCREAMING_SNAKE_CASE__ = BatchSampler(range(16 ) , batch_size=4 , drop_last=__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = SkipBatchSampler(__lowerCamelCase , 2 ) self.assertListEqual(list(__lowerCamelCase ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowercase_ ( self : Union[str, Any] ) -> str: SCREAMING_SNAKE_CASE__ = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowercase_ ( self : Dict ) -> List[Any]: SCREAMING_SNAKE_CASE__ = DataLoader(list(range(16 ) ) , batch_size=4 ) SCREAMING_SNAKE_CASE__ = skip_first_batches(__lowerCamelCase , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def lowercase_ ( self : List[Any] ) -> List[str]: SCREAMING_SNAKE_CASE__ = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def lowercase_ ( self : Union[str, Any] ) -> str: Accelerator() SCREAMING_SNAKE_CASE__ = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
218
0
'''simple docstring''' import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=UpperCAmelCase__ ) class a__ ( UpperCAmelCase__ ): lowerCamelCase : str =field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCamelCase : ClassVar[Features] =Features({"image": Image()} ) lowerCamelCase : ClassVar[Features] =Features({"labels": ClassLabel} ) lowerCamelCase : str ="image" lowerCamelCase : str ="labels" def SCREAMING_SNAKE_CASE__ ( self : Any , a : int ): """simple docstring""" if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , a ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) __lowerCamelCase = copy.deepcopy(self ) __lowerCamelCase = self.label_schema.copy() __lowerCamelCase = features[self.label_column] __lowerCamelCase = label_schema return task_template @property def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
67
'''simple docstring''' import re from filelock import FileLock try: import nltk __UpperCAmelCase =True except (ImportError, ModuleNotFoundError): __UpperCAmelCase =False if NLTK_AVAILABLE: with FileLock(".lock") as lock: nltk.download("punkt", quiet=True) def __lowerCAmelCase ( UpperCamelCase__ ) -> str: re.sub('''<n>''' , '''''' , UpperCamelCase__ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(UpperCamelCase__ ) )
67
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) _UpperCamelCase = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
16
'''simple docstring''' class _A : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=None ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase : int = data __UpperCAmelCase : int = previous __UpperCAmelCase : Union[str, Any] = next_node def __str__( self ) -> str: '''simple docstring''' return f'{self.data}' def __A ( self ) -> int: '''simple docstring''' return self.data def __A ( self ) -> List[str]: '''simple docstring''' return self.next def __A ( self ) -> str: '''simple docstring''' return self.previous class _A : def __init__( self , __UpperCAmelCase ) -> str: '''simple docstring''' __UpperCAmelCase : int = head def __iter__( self ) -> str: '''simple docstring''' return self def __A ( self ) -> str: '''simple docstring''' if not self.current: raise StopIteration else: __UpperCAmelCase : List[str] = self.current.get_data() __UpperCAmelCase : int = self.current.get_next() return value class _A : def __init__( self ) -> List[Any]: '''simple docstring''' __UpperCAmelCase : Optional[Any] = None # First node in list __UpperCAmelCase : List[str] = None # Last node in list def __str__( self ) -> int: '''simple docstring''' __UpperCAmelCase : Tuple = self.head __UpperCAmelCase : Optional[int] = [] while current is not None: nodes.append(current.get_data() ) __UpperCAmelCase : Any = current.get_next() return " ".join(str(__UpperCAmelCase ) for node in nodes ) def __contains__( self , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase : List[Any] = self.head while current: if current.get_data() == value: return True __UpperCAmelCase : Optional[Any] = current.get_next() return False def __iter__( self ) -> str: '''simple docstring''' return LinkedListIterator(self.head ) def __A ( self ) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def __A ( self ) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def __A ( self , __UpperCAmelCase ) -> None: '''simple docstring''' if self.head is None: __UpperCAmelCase : str = node __UpperCAmelCase : List[str] = node else: self.insert_before_node(self.head , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase ) -> None: '''simple docstring''' if self.head is None: self.set_head(__UpperCAmelCase ) else: self.insert_after_node(self.tail , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : Optional[int] = Node(__UpperCAmelCase ) if self.head is None: self.set_head(__UpperCAmelCase ) else: self.set_tail(__UpperCAmelCase ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : Tuple = node __UpperCAmelCase : List[Any] = node.previous if node.get_previous() is None: __UpperCAmelCase : str = node_to_insert else: __UpperCAmelCase : Optional[Any] = node_to_insert __UpperCAmelCase : List[Any] = node_to_insert def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : List[str] = node __UpperCAmelCase : Union[str, Any] = node.next if node.get_next() is None: __UpperCAmelCase : Dict = node_to_insert else: __UpperCAmelCase : Any = node_to_insert __UpperCAmelCase : List[str] = node_to_insert def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> None: '''simple docstring''' __UpperCAmelCase : Optional[Any] = 1 __UpperCAmelCase : Optional[Any] = Node(__UpperCAmelCase ) __UpperCAmelCase : Optional[Any] = self.head while node: if current_position == position: self.insert_before_node(__UpperCAmelCase , __UpperCAmelCase ) return current_position += 1 __UpperCAmelCase : int = node.next self.insert_after_node(self.tail , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase ) -> Node: '''simple docstring''' __UpperCAmelCase : Dict = self.head while node: if node.get_data() == item: return node __UpperCAmelCase : List[str] = node.get_next() raise Exception("""Node not found""" ) def __A ( self , __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' if (node := self.get_node(__UpperCAmelCase )) is not None: if node == self.head: __UpperCAmelCase : Optional[int] = self.head.get_next() if node == self.tail: __UpperCAmelCase : Union[str, Any] = self.tail.get_previous() self.remove_node_pointers(__UpperCAmelCase ) @staticmethod def __A ( __UpperCAmelCase ) -> None: '''simple docstring''' if node.get_next(): __UpperCAmelCase : Optional[Any] = node.previous if node.get_previous(): __UpperCAmelCase : int = node.next __UpperCAmelCase : Tuple = None __UpperCAmelCase : Union[str, Any] = None def __A ( self ) -> List[Any]: '''simple docstring''' return self.head is None def lowercase_ ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
16
1
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase ): assert ( isinstance(__UpperCamelCase , __UpperCamelCase ) and number_of_steps > 0 ), F"""number_of_steps needs to be positive integer, your input {number_of_steps}""" if number_of_steps == 1: return 1 lowercase__ : Union[str, Any] = 1, 1 for _ in range(number_of_steps - 1 ): lowercase__ : Union[str, Any] = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
198
"""simple docstring""" import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowerCamelCase ( self : Optional[Any] ): lowerCAmelCase_ : Optional[int] = 1 lowerCAmelCase_ : int = 3 lowerCAmelCase_ : Dict = (32, 32) lowerCAmelCase_ : List[Any] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(a_ ) return image @property def lowerCamelCase ( self : List[Any] ): torch.manual_seed(0 ) lowerCAmelCase_ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def lowerCamelCase ( self : Tuple ): torch.manual_seed(0 ) lowerCAmelCase_ : Dict = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def lowerCamelCase ( self : List[str] ): torch.manual_seed(0 ) lowerCAmelCase_ : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(a_ ) @property def lowerCamelCase ( self : Union[str, Any] ): def extract(*a_ : Tuple , **a_ : Tuple ): class __lowerCamelCase : '''simple docstring''' def __init__( self : Union[str, Any] ): lowerCAmelCase_ : List[str] = torch.ones([0] ) def lowerCamelCase ( self : str , a_ : Optional[int] ): self.pixel_values.to(a_ ) return self return Out() return extract def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : List[Any] = self.dummy_cond_unet lowerCAmelCase_ : List[Any] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=a_ , set_alpha_to_one=a_ , ) lowerCAmelCase_ : List[Any] = self.dummy_vae lowerCAmelCase_ : List[str] = self.dummy_text_encoder lowerCAmelCase_ : List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) # make sure here that pndm scheduler skips prk lowerCAmelCase_ : Optional[Any] = StableDiffusionPipeline( unet=a_ , scheduler=a_ , vae=a_ , text_encoder=a_ , tokenizer=a_ , safety_checker=a_ , feature_extractor=self.dummy_extractor , ) lowerCAmelCase_ : Union[str, Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) lowerCAmelCase_ : str = "A painting of a squirrel eating a burger" lowerCAmelCase_ : Any = torch.Generator(device=a_ ).manual_seed(0 ) lowerCAmelCase_ : Union[str, Any] = sd_pipe([prompt] , generator=a_ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" ) lowerCAmelCase_ : str = output.images lowerCAmelCase_ : Dict = torch.Generator(device=a_ ).manual_seed(0 ) lowerCAmelCase_ : str = sd_pipe( [prompt] , generator=a_ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , return_dict=a_ , )[0] lowerCAmelCase_ : str = image[0, -3:, -3:, -1] lowerCAmelCase_ : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCAmelCase_ : Any = np.array([0.5756, 0.6118, 0.5005, 0.5041, 0.5471, 0.4726, 0.4976, 0.4865, 0.4864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : List[str] = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Union[str, Any] = self.dummy_cond_unet lowerCAmelCase_ : Any = PNDMScheduler(skip_prk_steps=a_ ) lowerCAmelCase_ : List[Any] = self.dummy_vae lowerCAmelCase_ : List[str] = self.dummy_text_encoder lowerCAmelCase_ : Optional[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) # make sure here that pndm scheduler skips prk lowerCAmelCase_ : List[str] = StableDiffusionPipeline( unet=a_ , scheduler=a_ , vae=a_ , text_encoder=a_ , tokenizer=a_ , safety_checker=a_ , feature_extractor=self.dummy_extractor , ) lowerCAmelCase_ : Optional[Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) lowerCAmelCase_ : Optional[Any] = "A painting of a squirrel eating a burger" lowerCAmelCase_ : List[Any] = torch.Generator(device=a_ ).manual_seed(0 ) lowerCAmelCase_ : Any = sd_pipe([prompt] , generator=a_ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" ) lowerCAmelCase_ : Union[str, Any] = output.images lowerCAmelCase_ : List[str] = torch.Generator(device=a_ ).manual_seed(0 ) lowerCAmelCase_ : Optional[int] = sd_pipe( [prompt] , generator=a_ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , return_dict=a_ , )[0] lowerCAmelCase_ : Dict = image[0, -3:, -3:, -1] lowerCAmelCase_ : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCAmelCase_ : str = np.array([0.5125, 0.5716, 0.4828, 0.5060, 0.5650, 0.4768, 0.5185, 0.4895, 0.4993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : Tuple = StableDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-lms-pipe" , safety_checker=a_ ) assert isinstance(a_ , a_ ) assert isinstance(pipe.scheduler , a_ ) assert pipe.safety_checker is None lowerCAmelCase_ : str = pipe("example prompt" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(a_ ) lowerCAmelCase_ : List[str] = StableDiffusionPipeline.from_pretrained(a_ ) # sanity check that the pipeline still works assert pipe.safety_checker is None lowerCAmelCase_ : Any = pipe("example prompt" , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : str = self.dummy_cond_unet lowerCAmelCase_ : str = PNDMScheduler(skip_prk_steps=a_ ) lowerCAmelCase_ : Tuple = self.dummy_vae lowerCAmelCase_ : Dict = self.dummy_text_encoder lowerCAmelCase_ : Tuple = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) # put models in fp16 lowerCAmelCase_ : int = unet.half() lowerCAmelCase_ : Dict = vae.half() lowerCAmelCase_ : List[Any] = bert.half() # make sure here that pndm scheduler skips prk lowerCAmelCase_ : Optional[int] = StableDiffusionPipeline( unet=a_ , scheduler=a_ , vae=a_ , text_encoder=a_ , tokenizer=a_ , safety_checker=a_ , feature_extractor=self.dummy_extractor , ) lowerCAmelCase_ : Optional[Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) lowerCAmelCase_ : List[str] = "A painting of a squirrel eating a burger" lowerCAmelCase_ : Optional[Any] = sd_pipe([prompt] , num_inference_steps=2 , output_type="np" ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase ( self : Optional[int] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : List[str] ): lowerCAmelCase_ : Union[str, Any] = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5" , safety_checker=a_ ) lowerCAmelCase_ : Any = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowerCAmelCase_ : str = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) lowerCAmelCase_ : List[str] = ( "portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle" " coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with" " anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and" " children from bahnhof zoo, detailed " ) lowerCAmelCase_ : Optional[int] = 40_03_66_03_46 lowerCAmelCase_ : List[Any] = 7 # without safety guidance (sld_guidance_scale = 0) lowerCAmelCase_ : Union[str, Any] = torch.manual_seed(a_ ) lowerCAmelCase_ : Union[str, Any] = sd_pipe( [prompt] , generator=a_ , guidance_scale=a_ , num_inference_steps=50 , output_type="np" , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) lowerCAmelCase_ : Union[str, Any] = output.images lowerCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1] lowerCAmelCase_ : Union[str, Any] = [0.2278, 0.2231, 0.2249, 0.2333, 0.2303, 0.1885, 0.2273, 0.2144, 0.2176] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 # without safety guidance (strong configuration) lowerCAmelCase_ : List[str] = torch.manual_seed(a_ ) lowerCAmelCase_ : Any = sd_pipe( [prompt] , generator=a_ , guidance_scale=a_ , num_inference_steps=50 , output_type="np" , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowerCAmelCase_ : Optional[Any] = output.images lowerCAmelCase_ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase_ : Optional[Any] = [0.2383, 0.2276, 0.236, 0.2192, 0.2186, 0.2053, 0.1971, 0.1901, 0.1719] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCamelCase ( self : str ): lowerCAmelCase_ : Any = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5" , safety_checker=a_ ) lowerCAmelCase_ : Union[str, Any] = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) lowerCAmelCase_ : Optional[Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) lowerCAmelCase_ : Union[str, Any] = "padme amidala taking a bath artwork, safe for work, no nudity" lowerCAmelCase_ : Union[str, Any] = 27_34_97_17_55 lowerCAmelCase_ : Union[str, Any] = 7 lowerCAmelCase_ : str = torch.manual_seed(a_ ) lowerCAmelCase_ : Dict = sd_pipe( [prompt] , generator=a_ , guidance_scale=a_ , num_inference_steps=50 , output_type="np" , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) lowerCAmelCase_ : Any = output.images lowerCAmelCase_ : int = image[0, -3:, -3:, -1] lowerCAmelCase_ : int = [0.3502, 0.3622, 0.3396, 0.3642, 0.3478, 0.3318, 0.35, 0.3348, 0.3297] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 lowerCAmelCase_ : Optional[int] = torch.manual_seed(a_ ) lowerCAmelCase_ : Union[str, Any] = sd_pipe( [prompt] , generator=a_ , guidance_scale=a_ , num_inference_steps=50 , output_type="np" , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowerCAmelCase_ : Any = output.images lowerCAmelCase_ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = [0.5531, 0.5206, 0.4895, 0.5156, 0.5182, 0.4751, 0.4802, 0.4803, 0.4443] assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCamelCase ( self : Union[str, Any] ): lowerCAmelCase_ : Dict = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5" ) lowerCAmelCase_ : Any = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) lowerCAmelCase_ : Tuple = ( "the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c." " leyendecker" ) lowerCAmelCase_ : List[Any] = 10_44_35_52_34 lowerCAmelCase_ : Dict = 12 lowerCAmelCase_ : int = torch.manual_seed(a_ ) lowerCAmelCase_ : List[str] = sd_pipe( [prompt] , generator=a_ , guidance_scale=a_ , num_inference_steps=50 , output_type="np" , width=5_12 , height=5_12 , sld_guidance_scale=0 , ) lowerCAmelCase_ : int = output.images lowerCAmelCase_ : int = image[0, -3:, -3:, -1] lowerCAmelCase_ : int = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-7 lowerCAmelCase_ : int = torch.manual_seed(a_ ) lowerCAmelCase_ : Any = sd_pipe( [prompt] , generator=a_ , guidance_scale=a_ , num_inference_steps=50 , output_type="np" , width=5_12 , height=5_12 , sld_guidance_scale=20_00 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) lowerCAmelCase_ : Optional[Any] = output.images lowerCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1] lowerCAmelCase_ : str = np.array([0.5818, 0.6285, 0.6835, 0.6019, 0.625, 0.6754, 0.6096, 0.6334, 0.6561] ) assert image.shape == (1, 5_12, 5_12, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
241
0
"""simple docstring""" def lowercase__( __SCREAMING_SNAKE_CASE : int = 4_00_00_00 ): lowercase_ : str = [] lowercase_ , lowercase_ : List[Any] = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__SCREAMING_SNAKE_CASE ) lowercase_ , lowercase_ : List[str] = b, a + b return sum(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F"{solution() = }")
321
"""simple docstring""" from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
321
1
import math import random def lowercase__ ( __snake_case : float , __snake_case : bool = False ): '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value __UpperCAmelCase = 0.0_2 def lowercase__ ( __snake_case : int , __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : str = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(__snake_case ): # Forward propagation UpperCAmelCase_ : str = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCAmelCase_ : Union[str, Any] = (expected / 100) - layer_a # Error delta UpperCAmelCase_ : Any = layer_1_error * sigmoid_function(__snake_case , __snake_case ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase = int(input('Expected value: ')) __UpperCAmelCase = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
29
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = { 'configuration_time_series_transformer': [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimeSeriesTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TimeSeriesTransformerForPrediction', 'TimeSeriesTransformerModel', 'TimeSeriesTransformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
29
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __snake_case : str = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : List[Any] = ["""GPTSw3Tokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys __snake_case : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
365
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE : def __init__( self , _UpperCamelCase , _UpperCamelCase=3 , _UpperCamelCase=32 , _UpperCamelCase=3 , _UpperCamelCase=10 , _UpperCamelCase=[10, 20, 30, 40] , _UpperCamelCase=[1, 1, 2, 1] , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase="relu" , _UpperCamelCase=3 , _UpperCamelCase=None , ): """simple docstring""" lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = image_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = embeddings_size lowerCAmelCase__ = hidden_sizes lowerCAmelCase__ = depths lowerCAmelCase__ = is_training lowerCAmelCase__ = use_labels lowerCAmelCase__ = hidden_act lowerCAmelCase__ = num_labels lowerCAmelCase__ = scope lowerCAmelCase__ = len(_UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ = None if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase__ = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = TFResNetModel(config=_UpperCamelCase ) lowerCAmelCase__ = model(_UpperCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = self.num_labels lowerCAmelCase__ = TFResNetForImageClassification(_UpperCamelCase ) lowerCAmelCase__ = model(_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = config_and_inputs lowerCAmelCase__ = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class __SCREAMING_SNAKE_CASE ( __lowercase , __lowercase , unittest.TestCase): _SCREAMING_SNAKE_CASE : Any = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () _SCREAMING_SNAKE_CASE : List[str] = ( {'''feature-extraction''': TFResNetModel, '''image-classification''': TFResNetForImageClassification} if is_tf_available() else {} ) _SCREAMING_SNAKE_CASE : Union[str, Any] = False _SCREAMING_SNAKE_CASE : int = False _SCREAMING_SNAKE_CASE : Optional[int] = False _SCREAMING_SNAKE_CASE : List[str] = False _SCREAMING_SNAKE_CASE : Optional[Any] = False def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = TFResNetModelTester(self ) lowerCAmelCase__ = ConfigTester(self , config_class=_UpperCamelCase , has_text_modality=_UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase__ ( self ): """simple docstring""" return @unittest.skip(reason='ResNet does not use inputs_embeds' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip(reason='ResNet does not support input and output embeddings' ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(_UpperCamelCase ) lowerCAmelCase__ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ = [*signature.parameters.keys()] lowerCAmelCase__ = ['pixel_values'] self.assertListEqual(arg_names[:1] , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" def check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): lowerCAmelCase__ = model_class(_UpperCamelCase ) lowerCAmelCase__ = model(**self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) lowerCAmelCase__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase__ = self.model_tester.num_stages self.assertEqual(len(_UpperCamelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: lowerCAmelCase__ = layer_type lowerCAmelCase__ = True check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ = True check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_UpperCamelCase ) @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = TFResNetModel.from_pretrained(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) def _UpperCamelCase ( ) -> List[str]: """simple docstring""" lowerCAmelCase__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase): @cached_property def UpperCamelCase__ ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) lowerCAmelCase__ = self.default_image_processor lowerCAmelCase__ = prepare_img() lowerCAmelCase__ = image_processor(images=_UpperCamelCase , return_tensors='tf' ) # forward pass lowerCAmelCase__ = model(**_UpperCamelCase ) # verify the logits lowerCAmelCase__ = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , _UpperCamelCase ) lowerCAmelCase__ = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , _UpperCamelCase , atol=1E-4 ) )
122
0
import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('''0.12.2'''): raise Exception('''requires fairseq >= 0.12.2''') if version.parse(fairseq.__version__) > version.parse('''2'''): raise Exception('''requires fairseq < v2''') logging.set_verbosity_info() UpperCAmelCase : List[str] = logging.get_logger(__name__) UpperCAmelCase : List[str] = '''Hello, World!''' UpperCAmelCase : str = '''en_XX''' def _SCREAMING_SNAKE_CASE ( a , a , a ) -> Optional[int]: __A : Any = Path('data_bin' ) __A : Any = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(a ).parent ) , checkpoint_file=Path(a ).name , _name='xmod_base' , arch='xmod_base' , task='multilingual_masked_lm' , data_name_or_path=str(a ) , bpe='sentencepiece' , sentencepiece_model=str(Path(a ).parent / 'sentencepiece.bpe.model' ) , src_dict=str(data_dir / 'dict.txt' ) , ) xmod.eval() # disable dropout print(a ) __A : Dict = xmod.model.encoder.sentence_encoder __A : str = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings , hidden_size=xmod.cfg.model.encoder_embed_dim , num_hidden_layers=xmod.cfg.model.encoder_layers , num_attention_heads=xmod.cfg.model.encoder_attention_heads , intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=5_14 , type_vocab_size=1 , layer_norm_eps=1e-5 , pre_norm=xmod.cfg.model.encoder_normalize_before , adapter_reduction_factor=getattr(xmod.cfg.model , 'bottleneck' , 2 ) , adapter_layer_norm=xmod.cfg.model.adapter_layer_norm , adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm , ln_before_adapter=xmod.cfg.model.ln_before_adapter , languages=xmod.cfg.model.languages , ) if classification_head: __A : Tuple = xmod.model.classification_heads['mnli'].out_proj.weight.shape[0] print('Our X-MOD config:' , a ) __A : Any = XmodForSequenceClassification(a ) if classification_head else XmodForMaskedLM(a ) model.eval() # Now let's copy all the weights. # Embeddings __A : str = xmod_sent_encoder.embed_tokens.weight __A : Optional[Any] = xmod_sent_encoder.embed_positions.weight __A : Any = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. __A : Dict = xmod_sent_encoder.layernorm_embedding.weight __A : List[str] = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer __A : Optional[int] = model.roberta.encoder.layer[i] __A : Optional[int] = xmod_sent_encoder.layers[i] # self attention __A : Union[str, Any] = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError('Dimensions of self-attention weights do not match.' ) __A : Tuple = xmod_layer.self_attn.q_proj.weight __A : Optional[Any] = xmod_layer.self_attn.q_proj.bias __A : Optional[int] = xmod_layer.self_attn.k_proj.weight __A : str = xmod_layer.self_attn.k_proj.bias __A : int = xmod_layer.self_attn.v_proj.weight __A : Optional[Any] = xmod_layer.self_attn.v_proj.bias # self-attention output __A : Dict = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError('Dimensions of self-attention output weights do not match.' ) __A : int = xmod_layer.self_attn.out_proj.weight __A : Dict = xmod_layer.self_attn.out_proj.bias __A : List[str] = xmod_layer.self_attn_layer_norm.weight __A : str = xmod_layer.self_attn_layer_norm.bias # intermediate __A : Optional[Any] = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('Dimensions of intermediate weights do not match.' ) __A : Any = xmod_layer.fca.weight __A : Tuple = xmod_layer.fca.bias # output __A : List[Any] = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('Dimensions of feed-forward weights do not match.' ) __A : str = xmod_layer.fca.weight __A : Tuple = xmod_layer.fca.bias __A : List[Any] = xmod_layer.final_layer_norm.weight __A : int = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: __A : List[str] = xmod_layer.adapter_layer_norm.weight __A : Dict = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError('Lists of language adapters do not match.' ) for lang_code, adapter in xmod_layer.adapter_modules.items(): __A : str = bert_output.adapter_modules[lang_code] __A : List[str] = xmod_layer.adapter_modules[lang_code] __A : List[str] = from_adapter.fca.weight __A : Tuple = from_adapter.fca.bias __A : Optional[int] = from_adapter.fca.weight __A : int = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: __A : List[Any] = xmod_sent_encoder.layer_norm.weight __A : str = xmod_sent_encoder.layer_norm.bias if classification_head: __A : Any = xmod.model.classification_heads['mnli'].dense.weight __A : Dict = xmod.model.classification_heads['mnli'].dense.bias __A : Any = xmod.model.classification_heads['mnli'].out_proj.weight __A : Any = xmod.model.classification_heads['mnli'].out_proj.bias else: # LM Head __A : List[Any] = xmod.model.encoder.lm_head.dense.weight __A : str = xmod.model.encoder.lm_head.dense.bias __A : Optional[int] = xmod.model.encoder.lm_head.layer_norm.weight __A : Union[str, Any] = xmod.model.encoder.lm_head.layer_norm.bias __A : Dict = xmod.model.encoder.lm_head.weight __A : int = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. __A : List[str] = xmod.encode(a ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(a ) __A : Optional[Any] = model(a )[0] if classification_head: __A : Optional[int] = xmod.model.classification_heads['mnli'](xmod.extract_features(a ) ) else: __A : Any = xmod.model(a , lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape , their_output.shape ) __A : int = torch.max(torch.abs(our_output - their_output ) ).item() print(F"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 __A : Union[str, Any] = torch.allclose(a , a , atol=1e-3 ) print('Do both models output the same tensors?' , '🔥' if success else '💩' ) if not success: raise Exception('Something went wRoNg' ) Path(a ).mkdir(parents=a , exist_ok=a ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(a ) if __name__ == "__main__": UpperCAmelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xmod_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--classification_head''', action='''store_true''', help='''Whether to convert a final classification head.''' ) UpperCAmelCase : Any = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
280
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase : Dict = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def _SCREAMING_SNAKE_CASE ( a , a , a=8 ) -> Tuple: __A : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __A : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def _SCREAMING_SNAKE_CASE ( a , a=5_12 , a=5_12 ) -> int: __A : Optional[Any] = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) __A : Union[str, Any] = np.array(pil_image.convert('RGB' ) ) __A : Optional[int] = arr.astype(np.floataa ) / 127.5 - 1 __A : int = np.transpose(a , [2, 0, 1] ) __A : Tuple = torch.from_numpy(a ).unsqueeze(0 ) return image class _A( snake_case__ ): """simple docstring""" def __init__( self , _A , _A , _A , ): super().__init__() self.register_modules( unet=_A , scheduler=_A , movq=_A , ) __A : Tuple = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCAmelCase_ ( self , _A , _A , _A ): # get the original timestep using init_timestep __A : Optional[int] = min(int(num_inference_steps * strength ) , _A ) __A : Dict = max(num_inference_steps - init_timestep , 0 ) __A : Tuple = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCAmelCase_ ( self , _A , _A , _A , _A , _A , _A , _A=None ): if not isinstance(_A , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_A )}""" ) __A : Union[str, Any] = image.to(device=_A , dtype=_A ) __A : Optional[Any] = batch_size * num_images_per_prompt if image.shape[1] == 4: __A : int = image else: if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(_A )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(_A , _A ): __A : str = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_A ) ] __A : str = torch.cat(_A , dim=0 ) else: __A : List[str] = self.movq.encode(_A ).latent_dist.sample(_A ) __A : Tuple = self.movq.config.scaling_factor * init_latents __A : Optional[int] = torch.cat([init_latents] , dim=0 ) __A : Union[str, Any] = init_latents.shape __A : List[str] = randn_tensor(_A , generator=_A , device=_A , dtype=_A ) # get latents __A : Optional[Any] = self.scheduler.add_noise(_A , _A , _A ) __A : Optional[int] = init_latents return latents def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) __A : Optional[int] = torch.device(F"""cuda:{gpu_id}""" ) __A : Union[str, Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_A , _A ) def UpperCAmelCase_ ( self , _A=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) __A : List[Any] = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __A : int = None for cpu_offloaded_model in [self.unet, self.movq]: __A , __A : Optional[int] = cpu_offload_with_hook(_A , _A , prev_module_hook=_A ) # We'll offload the last model manually. __A : List[str] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase_ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_A , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_A ) def __call__( self , _A , _A , _A , _A = 512 , _A = 512 , _A = 100 , _A = 4.0 , _A = 0.3 , _A = 1 , _A = None , _A = "pil" , _A = True , ): __A : List[Any] = self._execution_device __A : Optional[Any] = guidance_scale > 1.0 if isinstance(_A , _A ): __A : Optional[Any] = torch.cat(_A , dim=0 ) __A : Tuple = image_embeds.shape[0] if isinstance(_A , _A ): __A : List[Any] = torch.cat(_A , dim=0 ) if do_classifier_free_guidance: __A : Union[str, Any] = image_embeds.repeat_interleave(_A , dim=0 ) __A : Optional[int] = negative_image_embeds.repeat_interleave(_A , dim=0 ) __A : List[str] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_A ) if not isinstance(_A , _A ): __A : List[Any] = [image] if not all(isinstance(_A , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( F"""Input is in incorrect format: {[type(_A ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) __A : Dict = torch.cat([prepare_image(_A , _A , _A ) for i in image] , dim=0 ) __A : Any = image.to(dtype=image_embeds.dtype , device=_A ) __A : Tuple = self.movq.encode(_A )['latents'] __A : int = latents.repeat_interleave(_A , dim=0 ) self.scheduler.set_timesteps(_A , device=_A ) __A , __A : int = self.get_timesteps(_A , _A , _A ) __A : Union[str, Any] = timesteps[:1].repeat(batch_size * num_images_per_prompt ) __A , __A : Any = downscale_height_and_width(_A , _A , self.movq_scale_factor ) __A : Tuple = self.prepare_latents( _A , _A , _A , _A , image_embeds.dtype , _A , _A ) for i, t in enumerate(self.progress_bar(_A ) ): # expand the latents if we are doing classifier free guidance __A : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __A : Dict = {'image_embeds': image_embeds} __A : List[str] = self.unet( sample=_A , timestep=_A , encoder_hidden_states=_A , added_cond_kwargs=_A , return_dict=_A , )[0] if do_classifier_free_guidance: __A , __A : Dict = noise_pred.split(latents.shape[1] , dim=1 ) __A , __A : Optional[Any] = noise_pred.chunk(2 ) __A , __A : List[str] = variance_pred.chunk(2 ) __A : str = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __A : List[str] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __A , __A : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __A : List[str] = self.scheduler.step( _A , _A , _A , generator=_A , )[0] # post-processing __A : List[Any] = self.movq.decode(_A , force_not_quantize=_A )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: __A : List[str] = image * 0.5 + 0.5 __A : List[str] = image.clamp(0 , 1 ) __A : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __A : Any = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
280
1
'''simple docstring''' from __future__ import annotations import numpy as np def A__ ( UpperCAmelCase_ ): _UpperCamelCase , _UpperCamelCase : Optional[int] = np.shape(UpperCAmelCase_ ) if rows != columns: _UpperCamelCase : Union[str, Any] = ( '\'table\' has to be of square shaped array but got a ' f'{rows}x{columns} array:\n{table}' ) raise ValueError(UpperCAmelCase_ ) _UpperCamelCase : Optional[Any] = np.zeros((rows, columns) ) _UpperCamelCase : Tuple = np.zeros((rows, columns) ) for i in range(UpperCAmelCase_ ): for j in range(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = sum(lower[i][k] * upper[k][j] for k in range(UpperCAmelCase_ ) ) if upper[j][j] == 0: raise ArithmeticError('No LU decomposition exists' ) _UpperCamelCase : Optional[Any] = (table[i][j] - total) / upper[j][j] _UpperCamelCase : int = 1 for j in range(UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Optional[int] = sum(lower[i][k] * upper[k][j] for k in range(UpperCAmelCase_ ) ) _UpperCamelCase : Tuple = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
236
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = DanceDiffusionPipeline lowercase__ = UNCONDITIONAL_AUDIO_GENERATION_PARAMS lowercase__ = PipelineTesterMixin.required_optional_params - { """callback""", """latents""", """callback_steps""", """output_type""", """num_images_per_prompt""", } lowercase__ = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS lowercase__ = False lowercase__ = False def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _UpperCamelCase : str = UNetaDModel( block_out_channels=(32, 32, 64) ,extra_in_channels=16 ,sample_size=512 ,sample_rate=16000 ,in_channels=2 ,out_channels=2 ,flip_sin_to_cos=lowerCamelCase__ ,use_timestep_embedding=lowerCamelCase__ ,time_embedding_type='fourier' ,mid_block_type='UNetMidBlock1D' ,down_block_types=('DownBlock1DNoSkip', 'DownBlock1D', 'AttnDownBlock1D') ,up_block_types=('AttnUpBlock1D', 'UpBlock1D', 'UpBlock1DNoSkip') ,) _UpperCamelCase : int = IPNDMScheduler() _UpperCamelCase : List[str] = { 'unet': unet, 'scheduler': scheduler, } return components def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : int=0 ): '''simple docstring''' if str(lowerCamelCase__ ).startswith('mps' ): _UpperCamelCase : Union[str, Any] = torch.manual_seed(lowerCamelCase__ ) else: _UpperCamelCase : str = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _UpperCamelCase : str = { 'batch_size': 1, 'generator': generator, 'num_inference_steps': 4, } return inputs def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : List[str] = self.get_dummy_components() _UpperCamelCase : int = DanceDiffusionPipeline(**lowerCamelCase__ ) _UpperCamelCase : List[str] = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = self.get_dummy_inputs(lowerCamelCase__ ) _UpperCamelCase : List[Any] = pipe(**lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = output.audios _UpperCamelCase : List[str] = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) _UpperCamelCase : Dict = np.array([-0.7_2_6_5, 1.0_0_0_0, -0.8_3_8_8, 0.1_1_7_5, 0.9_4_9_8, -1.0_0_0_0] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def UpperCamelCase_ ( self : str ): '''simple docstring''' return super().test_save_load_local() @skip_mps def UpperCamelCase_ ( self : Dict ): '''simple docstring''' return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) @skip_mps def UpperCamelCase_ ( self : str ): '''simple docstring''' return super().test_save_load_optional_components() @skip_mps def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' return super().test_attention_slicing_forward_pass() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = torch_device _UpperCamelCase : Dict = DanceDiffusionPipeline.from_pretrained('harmonai/maestro-150k' ) _UpperCamelCase : str = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : int = torch.manual_seed(0 ) _UpperCamelCase : Optional[Any] = pipe(generator=lowerCamelCase__ ,num_inference_steps=100 ,audio_length_in_s=4.0_9_6 ) _UpperCamelCase : Optional[int] = output.audios _UpperCamelCase : int = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _UpperCamelCase : Tuple = np.array([-0.0_1_9_2, -0.0_2_3_1, -0.0_3_1_8, -0.0_0_5_9, 0.0_0_0_2, -0.0_0_2_0] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : List[Any] = torch_device _UpperCamelCase : int = DanceDiffusionPipeline.from_pretrained('harmonai/maestro-150k' ,torch_dtype=torch.floataa ) _UpperCamelCase : Any = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = torch.manual_seed(0 ) _UpperCamelCase : Union[str, Any] = pipe(generator=lowerCamelCase__ ,num_inference_steps=100 ,audio_length_in_s=4.0_9_6 ) _UpperCamelCase : Any = output.audios _UpperCamelCase : str = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _UpperCamelCase : Any = np.array([-0.0_3_6_7, -0.0_4_8_8, -0.0_7_7_1, -0.0_5_2_5, -0.0_4_4_4, -0.0_3_4_1] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2
236
1
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __lowerCamelCase : Dict = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.8.0''', '''To fix: pip install -r examples/pytorch/image-classification/requirements.txt''') __lowerCamelCase : Dict = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) __lowerCamelCase : int = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def _snake_case ( lowerCAmelCase : str ): """simple docstring""" with open(lowerCAmelCase , "rb" ) as f: SCREAMING_SNAKE_CASE_ : Union[str, Any] = Image.open(lowerCAmelCase ) return im.convert("RGB" ) @dataclass class a__ : A = field( default=A__ , metadata={ 'help': 'Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub).' } , ) A = field( default=A__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) A = field(default=A__ , metadata={'help': 'A folder containing the training data.'} ) A = field(default=A__ , metadata={'help': 'A folder containing the validation data.'} ) A = field( default=0.15 , metadata={'help': 'Percent to split off of train for validation.'} ) A = field( default=A__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) A = field( default=A__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) def __UpperCamelCase ( self : Tuple ): """simple docstring""" if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( "You must specify either a dataset name from the hub or a train and/or validation directory." ) @dataclass class a__ : A = field( default='google/vit-base-patch16-224-in21k' , metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} , ) A = field( default=A__ , metadata={'help': 'If training from scratch, pass a model type from the list: ' + ', '.join(A__ )} , ) A = field( default=A__ , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) A = field( default=A__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from s3'} ) A = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) A = field(default=A__ , metadata={'help': 'Name or path of preprocessor config.'} ) A = field( default=A__ , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) A = field( default=A__ , metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'} , ) def _snake_case ( lowerCAmelCase : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = torch.stack([example["pixel_values"] for example in examples] ) SCREAMING_SNAKE_CASE_ : str = torch.tensor([example["labels"] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def _snake_case ( ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[int] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_image_classification" , lowerCAmelCase , lowerCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() SCREAMING_SNAKE_CASE_ : Any = training_args.get_process_log_level() logger.setLevel(lowerCAmelCase ) transformers.utils.logging.set_verbosity(lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. SCREAMING_SNAKE_CASE_ : str = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: SCREAMING_SNAKE_CASE_ : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: SCREAMING_SNAKE_CASE_ : str = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task="image-classification" , use_auth_token=True if model_args.use_auth_token else None , ) else: SCREAMING_SNAKE_CASE_ : str = {} if data_args.train_dir is not None: SCREAMING_SNAKE_CASE_ : Optional[Any] = os.path.join(data_args.train_dir , "**" ) if data_args.validation_dir is not None: SCREAMING_SNAKE_CASE_ : int = os.path.join(data_args.validation_dir , "**" ) SCREAMING_SNAKE_CASE_ : Dict = load_dataset( "imagefolder" , data_files=lowerCAmelCase , cache_dir=model_args.cache_dir , task="image-classification" , ) # If we don't have a validation split, split off a percentage of train as validation. SCREAMING_SNAKE_CASE_ : Optional[Any] = None if "validation" in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , lowerCAmelCase ) and data_args.train_val_split > 0.0: SCREAMING_SNAKE_CASE_ : str = dataset["train"].train_test_split(data_args.train_val_split ) SCREAMING_SNAKE_CASE_ : str = split["train"] SCREAMING_SNAKE_CASE_ : List[Any] = split["test"] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. SCREAMING_SNAKE_CASE_ : Any = dataset["train"].features["labels"].names SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[int] = {}, {} for i, label in enumerate(lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : Tuple = str(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Any = label # Load the accuracy metric from the datasets package SCREAMING_SNAKE_CASE_ : Tuple = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowerCAmelCase : Optional[Any] ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) SCREAMING_SNAKE_CASE_ : str = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(lowerCAmelCase ) , labelaid=lowerCAmelCase , idalabel=lowerCAmelCase , finetuning_task="image-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) SCREAMING_SNAKE_CASE_ : List[str] = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) SCREAMING_SNAKE_CASE_ : List[str] = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: SCREAMING_SNAKE_CASE_ : Tuple = image_processor.size["shortest_edge"] else: SCREAMING_SNAKE_CASE_ : List[str] = (image_processor.size["height"], image_processor.size["width"]) SCREAMING_SNAKE_CASE_ : List[str] = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) SCREAMING_SNAKE_CASE_ : Dict = Compose( [ RandomResizedCrop(lowerCAmelCase ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) SCREAMING_SNAKE_CASE_ : Optional[Any] = Compose( [ Resize(lowerCAmelCase ), CenterCrop(lowerCAmelCase ), ToTensor(), normalize, ] ) def train_transforms(lowerCAmelCase : Dict ): SCREAMING_SNAKE_CASE_ : str = [ _train_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"] ] return example_batch def val_transforms(lowerCAmelCase : Union[str, Any] ): SCREAMING_SNAKE_CASE_ : int = [_val_transforms(pil_img.convert("RGB" ) ) for pil_img in example_batch["image"]] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError("--do_train requires a train dataset" ) if data_args.max_train_samples is not None: SCREAMING_SNAKE_CASE_ : List[str] = ( dataset["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(lowerCAmelCase ) if training_args.do_eval: if "validation" not in dataset: raise ValueError("--do_eval requires a validation dataset" ) if data_args.max_eval_samples is not None: SCREAMING_SNAKE_CASE_ : List[Any] = ( dataset["validation"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(lowerCAmelCase ) # Initalize our trainer SCREAMING_SNAKE_CASE_ : Dict = Trainer( model=lowerCAmelCase , args=lowerCAmelCase , train_dataset=dataset["train"] if training_args.do_train else None , eval_dataset=dataset["validation"] if training_args.do_eval else None , compute_metrics=lowerCAmelCase , tokenizer=lowerCAmelCase , data_collator=lowerCAmelCase , ) # Training if training_args.do_train: SCREAMING_SNAKE_CASE_ : List[Any] = None if training_args.resume_from_checkpoint is not None: SCREAMING_SNAKE_CASE_ : Any = training_args.resume_from_checkpoint elif last_checkpoint is not None: SCREAMING_SNAKE_CASE_ : List[Any] = last_checkpoint SCREAMING_SNAKE_CASE_ : Any = trainer.train(resume_from_checkpoint=lowerCAmelCase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: SCREAMING_SNAKE_CASE_ : int = trainer.evaluate() trainer.log_metrics("eval" , lowerCAmelCase ) trainer.save_metrics("eval" , lowerCAmelCase ) # Write model card and (optionally) push to hub SCREAMING_SNAKE_CASE_ : Optional[int] = { "finetuned_from": model_args.model_name_or_path, "tasks": "image-classification", "dataset": data_args.dataset_name, "tags": ["image-classification", "vision"], } if training_args.push_to_hub: trainer.push_to_hub(**lowerCAmelCase ) else: trainer.create_model_card(**lowerCAmelCase ) if __name__ == "__main__": main()
18
from collections.abc import Sequence from queue import Queue class a__ : def __init__( self : int,_A : List[Any],_A : Optional[Any],_A : Optional[int],_A : int=None,_A : List[str]=None ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = start SCREAMING_SNAKE_CASE_ : List[str] = end SCREAMING_SNAKE_CASE_ : Tuple = val SCREAMING_SNAKE_CASE_ : List[str] = (start + end) // 2 SCREAMING_SNAKE_CASE_ : Optional[int] = left SCREAMING_SNAKE_CASE_ : str = right def __repr__( self : Tuple ): """simple docstring""" return F'SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})' class a__ : def __init__( self : Any,_A : Sequence,_A : str ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = collection SCREAMING_SNAKE_CASE_ : Optional[int] = function if self.collection: SCREAMING_SNAKE_CASE_ : List[str] = self._build_tree(0,len(_A ) - 1 ) def __UpperCamelCase ( self : int,_A : Any,_A : List[Any] ): """simple docstring""" self._update_tree(self.root,_A,_A ) def __UpperCamelCase ( self : str,_A : Any,_A : List[Any] ): """simple docstring""" return self._query_range(self.root,_A,_A ) def __UpperCamelCase ( self : Any,_A : Optional[int],_A : int ): """simple docstring""" if start == end: return SegmentTreeNode(_A,_A,self.collection[start] ) SCREAMING_SNAKE_CASE_ : List[Any] = (start + end) // 2 SCREAMING_SNAKE_CASE_ : Union[str, Any] = self._build_tree(_A,_A ) SCREAMING_SNAKE_CASE_ : Optional[int] = self._build_tree(mid + 1,_A ) return SegmentTreeNode(_A,_A,self.fn(left.val,right.val ),_A,_A ) def __UpperCamelCase ( self : int,_A : int,_A : Tuple,_A : Dict ): """simple docstring""" if node.start == i and node.end == i: SCREAMING_SNAKE_CASE_ : Union[str, Any] = val return if i <= node.mid: self._update_tree(node.left,_A,_A ) else: self._update_tree(node.right,_A,_A ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.fn(node.left.val,node.right.val ) def __UpperCamelCase ( self : str,_A : List[str],_A : Optional[int],_A : Optional[Any] ): """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left,_A,_A ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left,_A,node.mid ),self._query_range(node.right,node.mid + 1,_A ),) else: # range in right child tree return self._query_range(node.right,_A,_A ) def __UpperCamelCase ( self : Optional[Any] ): """simple docstring""" if self.root is not None: SCREAMING_SNAKE_CASE_ : int = Queue() queue.put(self.root ) while not queue.empty(): SCREAMING_SNAKE_CASE_ : Tuple = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('''*''' * 50) __lowerCamelCase : int = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
18
1
"""simple docstring""" import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class __lowerCAmelCase : @staticmethod def UpperCAmelCase ( *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' pass def A ( snake_case :Image ) -> str: __UpperCamelCase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class __lowerCAmelCase ( unittest.TestCase ): lowercase = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = DepthEstimationPipeline(model=__UpperCAmelCase , image_processor=__UpperCAmelCase ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = depth_estimator('./tests/fixtures/tests_samples/COCO/000000039769.png' ) self.assertEqual({'predicted_depth': ANY(torch.Tensor ), 'depth': ANY(Image.Image )} , __UpperCAmelCase ) import datasets __UpperCamelCase = datasets.load_dataset('hf-internal-testing/fixtures_image_utils' , 'image' , split='test' ) __UpperCamelCase = depth_estimator( [ Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ), 'http://images.cocodataset.org/val2017/000000039769.jpg', # RGBA dataset[0]['file'], # LA dataset[1]['file'], # L dataset[2]['file'], ] ) self.assertEqual( [ {'predicted_depth': ANY(torch.Tensor ), 'depth': ANY(Image.Image )}, {'predicted_depth': ANY(torch.Tensor ), 'depth': ANY(Image.Image )}, {'predicted_depth': ANY(torch.Tensor ), 'depth': ANY(Image.Image )}, {'predicted_depth': ANY(torch.Tensor ), 'depth': ANY(Image.Image )}, {'predicted_depth': ANY(torch.Tensor ), 'depth': ANY(Image.Image )}, ] , __UpperCAmelCase , ) @require_tf @unittest.skip('Depth estimation is not implemented in TF' ) def UpperCAmelCase ( self ): '''simple docstring''' pass @slow @require_torch def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = 'Intel/dpt-large' __UpperCamelCase = pipeline('depth-estimation' , model=__UpperCAmelCase ) __UpperCamelCase = depth_estimator('http://images.cocodataset.org/val2017/000000039769.jpg' ) __UpperCamelCase = hashimage(outputs['depth'] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['predicted_depth'].max().item() ) , 2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['predicted_depth'].min().item() ) , 2.6_6_2 ) @require_torch def UpperCAmelCase ( self ): '''simple docstring''' self.skipTest('There is not hf-internal-testing tiny model for either GLPN nor DPT' )
263
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def A ( snake_case :str , snake_case :str = "cpu" , snake_case :Union[str, None] = None ) -> None: __UpperCamelCase = torch.load(snake_case , map_location=snake_case ) for k, v in tqdm(state_dict.items() ): if not isinstance(snake_case , torch.Tensor ): raise TypeError('FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin' ) __UpperCamelCase = v.half() if save_path is None: # overwrite src_path __UpperCamelCase = src_path torch.save(snake_case , snake_case ) if __name__ == "__main__": fire.Fire(convert)
263
1
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE__ : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class UpperCamelCase__ (lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : str = XLMRobertaTokenizer lowerCamelCase_ : str = XLMRobertaTokenizerFast lowerCamelCase_ : int = True lowerCamelCase_ : List[Any] = True def _lowercase ( self ) -> Optional[int]: super().setUp() # We have a SentencePiece fixture for testing lowerCamelCase : Any = XLMRobertaTokenizer(UpperCamelCase__ , keep_accents=UpperCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self ) -> List[str]: lowerCamelCase : Optional[int] = "<pad>" lowerCamelCase : Union[str, Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCamelCase__ ) , UpperCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCamelCase__ ) , UpperCamelCase__ ) def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : str = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(UpperCamelCase__ ) , 1002 ) def _lowercase ( self ) -> Dict: self.assertEqual(self.get_tokenizer().vocab_size , 1002 ) def _lowercase ( self ) -> Any: lowerCamelCase : List[str] = XLMRobertaTokenizer(UpperCamelCase__ , keep_accents=UpperCamelCase__ ) lowerCamelCase : str = tokenizer.tokenize("This is a test" ) self.assertListEqual(UpperCamelCase__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCamelCase : int = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) lowerCamelCase : Optional[int] = tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCamelCase : Any = tokenizer.convert_ids_to_tokens(UpperCamelCase__ ) self.assertListEqual( UpperCamelCase__ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def _lowercase ( self ) -> Any: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCamelCase : Dict = (self.rust_tokenizer_class, "hf-internal-testing/tiny-xlm-roberta", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCamelCase : str = self.rust_tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase : Any = self.tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase : Optional[Any] = tempfile.mkdtemp() lowerCamelCase : Dict = tokenizer_r.save_pretrained(UpperCamelCase__ ) lowerCamelCase : Tuple = tokenizer_p.save_pretrained(UpperCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) lowerCamelCase : Any = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(UpperCamelCase__ , UpperCamelCase__ ) # Checks everything loads correctly in the same way lowerCamelCase : int = tokenizer_r.from_pretrained(UpperCamelCase__ ) lowerCamelCase : Any = tokenizer_p.from_pretrained(UpperCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase__ , UpperCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCamelCase__ ) # Save tokenizer rust, legacy_format=True lowerCamelCase : str = tempfile.mkdtemp() lowerCamelCase : int = tokenizer_r.save_pretrained(UpperCamelCase__ , legacy_format=UpperCamelCase__ ) lowerCamelCase : List[str] = tokenizer_p.save_pretrained(UpperCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase__ , UpperCamelCase__ ) # Checks everything loads correctly in the same way lowerCamelCase : Optional[int] = tokenizer_r.from_pretrained(UpperCamelCase__ ) lowerCamelCase : Any = tokenizer_p.from_pretrained(UpperCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase__ , UpperCamelCase__ ) ) shutil.rmtree(UpperCamelCase__ ) # Save tokenizer rust, legacy_format=False lowerCamelCase : List[str] = tempfile.mkdtemp() lowerCamelCase : Any = tokenizer_r.save_pretrained(UpperCamelCase__ , legacy_format=UpperCamelCase__ ) lowerCamelCase : int = tokenizer_p.save_pretrained(UpperCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCamelCase : List[str] = tokenizer_r.from_pretrained(UpperCamelCase__ ) lowerCamelCase : Optional[Any] = tokenizer_p.from_pretrained(UpperCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase__ , UpperCamelCase__ ) ) shutil.rmtree(UpperCamelCase__ ) @cached_property def _lowercase ( self ) -> List[str]: return XLMRobertaTokenizer.from_pretrained("xlm-roberta-base" ) def _lowercase ( self ) -> List[Any]: with tempfile.NamedTemporaryFile() as f: shutil.copyfile(UpperCamelCase__ , f.name ) lowerCamelCase : Optional[Any] = XLMRobertaTokenizer(f.name , keep_accents=UpperCamelCase__ ) lowerCamelCase : Optional[Any] = pickle.dumps(UpperCamelCase__ ) pickle.loads(UpperCamelCase__ ) def _lowercase ( self ) -> int: if not self.test_rust_tokenizer: return lowerCamelCase : Dict = self.get_tokenizer() lowerCamelCase : Optional[Any] = self.get_rust_tokenizer() lowerCamelCase : int = "I was born in 92000, and this is falsé." lowerCamelCase : List[str] = tokenizer.tokenize(UpperCamelCase__ ) lowerCamelCase : List[Any] = rust_tokenizer.tokenize(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase : str = tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) lowerCamelCase : Union[str, Any] = rust_tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) lowerCamelCase : Optional[Any] = self.get_rust_tokenizer() lowerCamelCase : List[str] = tokenizer.encode(UpperCamelCase__ ) lowerCamelCase : int = rust_tokenizer.encode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) @slow def _lowercase ( self ) -> Dict: lowerCamelCase : str = "Hello World!" lowerCamelCase : Optional[Any] = [0, 3_5378, 6661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(UpperCamelCase__ , self.big_tokenizer.encode(UpperCamelCase__ ) ) @slow def _lowercase ( self ) -> int: lowerCamelCase : List[str] = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) lowerCamelCase : Union[str, Any] = [ 0, 3293, 83, 10, 4552, 4989, 7986, 678, 10, 5915, 111, 17_9459, 12_4850, 4, 6044, 237, 12, 6, 5, 6, 4, 6780, 705, 15, 1388, 44, 378, 1_0114, 711, 152, 20, 6, 5, 2_2376, 642, 1221, 1_5190, 3_4153, 450, 5608, 959, 1119, 5_7702, 136, 186, 47, 1098, 2_9367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6044, 237, 6284, 5_0901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(UpperCamelCase__ , self.big_tokenizer.encode(UpperCamelCase__ ) ) @slow def _lowercase ( self ) -> Any: # fmt: off lowerCamelCase : int = {"input_ids": [[0, 1_1062, 8_2772, 7, 15, 8_2772, 538, 5_1529, 237, 1_7198, 1290, 206, 9, 21_5175, 1314, 136, 1_7198, 1290, 206, 9, 5_6359, 42, 12_2009, 9, 1_6466, 16, 8_7344, 4537, 9, 4717, 7_8381, 6, 15_9958, 7, 15, 2_4480, 618, 4, 527, 2_2693, 5428, 4, 2777, 2_4480, 9874, 4, 4_3523, 594, 4, 803, 1_8392, 3_3189, 18, 4, 4_3523, 2_4447, 1_2399, 100, 2_4955, 8_3658, 9626, 14_4057, 15, 839, 2_2335, 16, 136, 2_4955, 8_3658, 8_3479, 15, 3_9102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 12_2009, 11_5774, 23, 805, 1328, 4_6876, 7, 136, 5_3894, 1940, 4_2227, 4_1159, 1_7721, 823, 425, 4, 2_7512, 9_8722, 206, 136, 5531, 4970, 919, 1_7336, 5, 2], [0, 2_0080, 618, 83, 8_2775, 47, 479, 9, 1517, 73, 5_3894, 333, 8_0581, 11_0117, 1_8811, 5256, 1295, 51, 15_2526, 297, 7986, 390, 12_4416, 538, 3_5431, 214, 98, 1_5044, 2_5737, 136, 7108, 4_3701, 23, 756, 13_5355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 6_3773, 11_9455, 6, 14_7797, 8_8203, 7, 645, 70, 21, 3285, 1_0269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCamelCase__ , model_name="xlm-roberta-base" , revision="d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3" , )
48
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE__ : Optional[int] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Dict = { 'salesforce/blip2-opt-2.7b': 'https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json', } class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : Union[str, Any] = """blip_2_vision_model""" def __init__( self , UpperCamelCase__=1408 , UpperCamelCase__=6144 , UpperCamelCase__=39 , UpperCamelCase__=16 , UpperCamelCase__=224 , UpperCamelCase__=14 , UpperCamelCase__="gelu" , UpperCamelCase__=0.00001 , UpperCamelCase__=0.0 , UpperCamelCase__=1e-10 , UpperCamelCase__=True , **UpperCamelCase__ , ) -> Optional[Any]: super().__init__(**UpperCamelCase__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : Union[str, Any] = intermediate_size lowerCamelCase : List[str] = num_hidden_layers lowerCamelCase : List[str] = num_attention_heads lowerCamelCase : Dict = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Dict = initializer_range lowerCamelCase : Union[str, Any] = attention_dropout lowerCamelCase : Dict = layer_norm_eps lowerCamelCase : Optional[Any] = hidden_act lowerCamelCase : str = qkv_bias @classmethod def _lowercase ( cls , UpperCamelCase__ , **UpperCamelCase__ ) -> "PretrainedConfig": cls._set_token_in_kwargs(UpperCamelCase__ ) lowerCamelCase , lowerCamelCase : List[str] = cls.get_config_dict(UpperCamelCase__ , **UpperCamelCase__ ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": lowerCamelCase : Optional[int] = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(UpperCamelCase__ , **UpperCamelCase__ ) class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : Dict = """blip_2_qformer""" def __init__( self , UpperCamelCase__=3_0522 , UpperCamelCase__=768 , UpperCamelCase__=12 , UpperCamelCase__=12 , UpperCamelCase__=3072 , UpperCamelCase__="gelu" , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=512 , UpperCamelCase__=0.02 , UpperCamelCase__=1e-12 , UpperCamelCase__=0 , UpperCamelCase__="absolute" , UpperCamelCase__=2 , UpperCamelCase__=1408 , **UpperCamelCase__ , ) -> int: super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase : Optional[int] = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Dict = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : int = hidden_act lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : Dict = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : Dict = max_position_embeddings lowerCamelCase : List[str] = initializer_range lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = position_embedding_type lowerCamelCase : Tuple = cross_attention_frequency lowerCamelCase : Optional[int] = encoder_hidden_size @classmethod def _lowercase ( cls , UpperCamelCase__ , **UpperCamelCase__ ) -> "PretrainedConfig": cls._set_token_in_kwargs(UpperCamelCase__ ) lowerCamelCase , lowerCamelCase : str = cls.get_config_dict(UpperCamelCase__ , **UpperCamelCase__ ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": lowerCamelCase : int = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(UpperCamelCase__ , **UpperCamelCase__ ) class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : List[str] = """blip-2""" lowerCamelCase_ : int = True def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=32 , **UpperCamelCase__ ) -> str: super().__init__(**UpperCamelCase__ ) if vision_config is None: lowerCamelCase : List[Any] = {} logger.info("vision_config is None. initializing the Blip2VisionConfig with default values." ) if qformer_config is None: lowerCamelCase : List[Any] = {} logger.info("qformer_config is None. Initializing the Blip2QFormerConfig with default values." ) if text_config is None: lowerCamelCase : Any = {} logger.info("text_config is None. Initializing the text config with default values (`OPTConfig`)." ) lowerCamelCase : Optional[int] = BlipaVisionConfig(**UpperCamelCase__ ) lowerCamelCase : str = BlipaQFormerConfig(**UpperCamelCase__ ) lowerCamelCase : List[str] = text_config["model_type"] if "model_type" in text_config else "opt" lowerCamelCase : str = CONFIG_MAPPING[text_model_type](**UpperCamelCase__ ) lowerCamelCase : Optional[Any] = self.text_config.tie_word_embeddings lowerCamelCase : int = self.text_config.is_encoder_decoder lowerCamelCase : Optional[Any] = num_query_tokens lowerCamelCase : int = self.vision_config.hidden_size lowerCamelCase : Tuple = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES lowerCamelCase : Dict = 1.0 lowerCamelCase : List[Any] = 0.02 @classmethod def _lowercase ( cls , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ , ) -> str: return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **UpperCamelCase__ , ) def _lowercase ( self ) -> Optional[Any]: lowerCamelCase : Tuple = copy.deepcopy(self.__dict__ ) lowerCamelCase : Tuple = self.vision_config.to_dict() lowerCamelCase : int = self.qformer_config.to_dict() lowerCamelCase : Optional[Any] = self.text_config.to_dict() lowerCamelCase : int = self.__class__.model_type return output
48
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { 'microsoft/swin-tiny-patch4-window7-224': ( 'https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json' ), # See all Swin models at https://huggingface.co/models?filter=swin } class __A (snake_case__ , snake_case__): '''simple docstring''' __lowercase: Dict = """swin""" __lowercase: List[Any] = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : Tuple , UpperCAmelCase_ : List[str]=224 , UpperCAmelCase_ : List[str]=4 , UpperCAmelCase_ : Optional[Any]=3 , UpperCAmelCase_ : List[str]=96 , UpperCAmelCase_ : Any=[2, 2, 6, 2] , UpperCAmelCase_ : Union[str, Any]=[3, 6, 12, 24] , UpperCAmelCase_ : Any=7 , UpperCAmelCase_ : List[str]=4.0 , UpperCAmelCase_ : Optional[int]=True , UpperCAmelCase_ : Optional[int]=0.0 , UpperCAmelCase_ : int=0.0 , UpperCAmelCase_ : List[str]=0.1 , UpperCAmelCase_ : int="gelu" , UpperCAmelCase_ : Tuple=False , UpperCAmelCase_ : int=0.02 , UpperCAmelCase_ : Dict=1E-5 , UpperCAmelCase_ : str=32 , UpperCAmelCase_ : Dict=None , UpperCAmelCase_ : str=None , **UpperCAmelCase_ : Any , ) ->Optional[int]: """simple docstring""" super().__init__(**UpperCAmelCase_ ) snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = len(UpperCAmelCase_ ) snake_case_ = num_heads snake_case_ = window_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = use_absolute_embeddings snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model snake_case_ = int(embed_dim * 2 ** (len(UpperCAmelCase_ ) - 1) ) snake_case_ = ["""stem"""] + [F"""stage{idx}""" for idx in range(1 , len(UpperCAmelCase_ ) + 1 )] snake_case_ , snake_case_ = get_aligned_output_features_output_indices( out_features=UpperCAmelCase_ , out_indices=UpperCAmelCase_ , stage_names=self.stage_names ) class __A (snake_case__): '''simple docstring''' __lowercase: List[str] = version.parse("""1.11""") @property def lowerCAmelCase ( self : List[Any] ) ->Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCAmelCase ( self : str ) ->float: """simple docstring""" return 1E-4
233
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __A (snake_case__): '''simple docstring''' __lowercase: Any = ["""image_processor""", """tokenizer"""] __lowercase: Dict = """AutoImageProcessor""" __lowercase: List[Any] = """AutoTokenizer""" def __init__( self : Tuple , UpperCAmelCase_ : int=None , UpperCAmelCase_ : Tuple=None , **UpperCAmelCase_ : str ) ->Optional[int]: """simple docstring""" snake_case_ = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , UpperCAmelCase_ , ) snake_case_ = kwargs.pop("""feature_extractor""" ) snake_case_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(UpperCAmelCase_ , UpperCAmelCase_ ) snake_case_ = self.image_processor snake_case_ = False def __call__( self : Any , *UpperCAmelCase_ : Optional[int] , **UpperCAmelCase_ : str ) ->List[str]: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*UpperCAmelCase_ , **UpperCAmelCase_ ) snake_case_ = kwargs.pop("""images""" , UpperCAmelCase_ ) snake_case_ = kwargs.pop("""text""" , UpperCAmelCase_ ) if len(UpperCAmelCase_ ) > 0: snake_case_ = args[0] snake_case_ = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: snake_case_ = self.image_processor(UpperCAmelCase_ , *UpperCAmelCase_ , **UpperCAmelCase_ ) if text is not None: snake_case_ = self.tokenizer(UpperCAmelCase_ , **UpperCAmelCase_ ) if text is None: return inputs elif images is None: return encodings else: snake_case_ = encodings["""input_ids"""] return inputs def lowerCAmelCase ( self : Optional[Any] , *UpperCAmelCase_ : Optional[Any] , **UpperCAmelCase_ : Tuple ) ->Optional[int]: """simple docstring""" return self.tokenizer.batch_decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) def lowerCAmelCase ( self : Optional[int] , *UpperCAmelCase_ : Union[str, Any] , **UpperCAmelCase_ : List[Any] ) ->str: """simple docstring""" return self.tokenizer.decode(*UpperCAmelCase_ , **UpperCAmelCase_ ) @contextmanager def lowerCAmelCase ( self : List[str] ) ->Dict: """simple docstring""" warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) snake_case_ = True snake_case_ = self.tokenizer yield snake_case_ = self.image_processor snake_case_ = False def lowerCAmelCase ( self : int , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str=False , UpperCAmelCase_ : str=None ) ->Tuple: """simple docstring""" if added_vocab is None: snake_case_ = self.tokenizer.get_added_vocab() snake_case_ = {} while tokens: snake_case_ = re.search(R"""<s_(.*?)>""" , UpperCAmelCase_ , re.IGNORECASE ) if start_token is None: break snake_case_ = start_token.group(1 ) snake_case_ = re.search(RF"""</s_{key}>""" , UpperCAmelCase_ , re.IGNORECASE ) snake_case_ = start_token.group() if end_token is None: snake_case_ = tokens.replace(UpperCAmelCase_ , """""" ) else: snake_case_ = end_token.group() snake_case_ = re.escape(UpperCAmelCase_ ) snake_case_ = re.escape(UpperCAmelCase_ ) snake_case_ = re.search(F"""{start_token_escaped}(.*?){end_token_escaped}""" , UpperCAmelCase_ , re.IGNORECASE ) if content is not None: snake_case_ = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node snake_case_ = self.tokenajson(UpperCAmelCase_ , is_inner_value=UpperCAmelCase_ , added_vocab=UpperCAmelCase_ ) if value: if len(UpperCAmelCase_ ) == 1: snake_case_ = value[0] snake_case_ = value else: # leaf nodes snake_case_ = [] for leaf in content.split(R"""<sep/>""" ): snake_case_ = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": snake_case_ = leaf[1:-2] # for categorical special tokens output[key].append(UpperCAmelCase_ ) if len(output[key] ) == 1: snake_case_ = output[key][0] snake_case_ = tokens[tokens.find(UpperCAmelCase_ ) + len(UpperCAmelCase_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=UpperCAmelCase_ , added_vocab=UpperCAmelCase_ ) if len(UpperCAmelCase_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def lowerCAmelCase ( self : List[Any] ) ->Optional[int]: """simple docstring""" warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , UpperCAmelCase_ , ) return self.image_processor_class @property def lowerCAmelCase ( self : List[str] ) ->str: """simple docstring""" warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , UpperCAmelCase_ , ) return self.image_processor
233
1
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _a ( _lowerCAmelCase): _a : Dict = ['''image_processor''', '''tokenizer'''] _a : Optional[Any] = '''AutoImageProcessor''' _a : Any = '''AutoTokenizer''' def __init__( self : Tuple , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : str )-> Optional[Any]: super().__init__(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = self.image_processor def __call__( self : Union[str, Any] , _SCREAMING_SNAKE_CASE : int=None , _SCREAMING_SNAKE_CASE : List[Any]=None , _SCREAMING_SNAKE_CASE : List[str]=None , **_SCREAMING_SNAKE_CASE : Any )-> List[str]: if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: lowerCAmelCase__ : int = self.tokenizer(_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if images is not None: lowerCAmelCase__ : Any = self.image_processor(_SCREAMING_SNAKE_CASE , return_tensors=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if text is not None and images is not None: lowerCAmelCase__ : Any = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_SCREAMING_SNAKE_CASE ) , tensor_type=_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : List[Any] , *_SCREAMING_SNAKE_CASE : Optional[int] , **_SCREAMING_SNAKE_CASE : Union[str, Any] )-> Union[str, Any]: return self.tokenizer.batch_decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Union[str, Any] , *_SCREAMING_SNAKE_CASE : Dict , **_SCREAMING_SNAKE_CASE : str )-> Dict: return self.tokenizer.decode(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) @property def UpperCAmelCase__( self : Optional[Any] )-> str: return ["input_ids", "attention_mask", "pixel_values"]
131
'''simple docstring''' import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) lowerCAmelCase : Optional[int] =logging.getLogger() def UpperCAmelCase_ ( ): lowercase_ :Union[str, Any] = argparse.ArgumentParser() parser.add_argument("-f" ) lowercase_ :Any = parser.parse_args() return args.f class a_ ( _lowerCAmelCase ): def lowercase__ ( self : List[Any] ): """simple docstring""" lowercase_ :List[Any] = logging.StreamHandler(sys.stdout ) logger.addHandler(lowercase ) def lowercase__ ( self : str , lowercase : str ): """simple docstring""" lowercase_ :Dict = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowercase , "argv" , lowercase ): lowercase_ :int = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowercase , 0.6_66 ) @slow @require_torch_non_multi_gpu def lowercase__ ( self : Tuple ): """simple docstring""" lowercase_ :List[Any] = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowercase ) lowercase_ :List[str] = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase ) lowercase_ :Any = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase )
223
0
"""simple docstring""" import argparse import copy def _A ( _a : Optional[Any] ): """simple docstring""" A = {} with open(_a ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: A = [] _list.append([line.split()[1], line.split()[2]] ) A = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: A = [] _list.append([line.split()[0], line.split()[2]] ) A = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def _A ( _a : int , _a : str ): """simple docstring""" with open(_a ) as f: A = f.read(1 ) A = start_node A = [] A = start_node A = 0 while visiting not in first_solution: A = 1_0_0_0_0 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(_a ) and k[0] not in first_solution: A = k[1] A = k[0] first_solution.append(_a ) A = distance_of_first_solution + int(_a ) A = best_node first_solution.append(_a ) A = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 A = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0_0_0_0 ) return first_solution, distance_of_first_solution def _A ( _a : Any , _a : str ): """simple docstring""" A = [] for n in solution[1:-1]: A = solution.index(_a ) for kn in solution[1:-1]: A = solution.index(_a ) if n == kn: continue A = copy.deepcopy(_a ) A = kn A = n A = 0 for k in _tmp[:-1]: A = _tmp[_tmp.index(_a ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: A = distance + int(i[1] ) _tmp.append(_a ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) A = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda _a : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def _A ( _a : Any , _a : List[Any] , _a : Union[str, Any] , _a : List[str] , _a : List[str] ): """simple docstring""" A = 1 A = first_solution A = [] A = distance_of_first_solution A = solution while count <= iters: A = find_neighborhood(_a , _a ) A = 0 A = neighborhood[index_of_best_solution] A = len(_a ) - 1 A = False while not found: A = 0 while i < len(_a ): if best_solution[i] != solution[i]: A = best_solution[i] A = solution[i] break A = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) A = True A = best_solution[:-1] A = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: A = cost A = solution else: A = index_of_best_solution + 1 A = neighborhood[index_of_best_solution] if len(_a ) >= size: tabu_list.pop(0 ) A = count + 1 return best_solution_ever, best_cost def _A ( _a : Any=None ): """simple docstring""" A = generate_neighbours(args.File ) A , A = generate_first_solution( args.File , _a ) A , A = tabu_search( _a , _a , _a , args.Iterations , args.Size , ) print(f'Best solution: {best_sol}, with total distance: {best_cost}.' ) if __name__ == "__main__": UpperCAmelCase =argparse.ArgumentParser(description="Tabu Search") parser.add_argument( "-f", "--File", type=str, help="Path to the file containing the data", required=True, ) parser.add_argument( "-i", "--Iterations", type=int, help="How many iterations the algorithm should perform", required=True, ) parser.add_argument( "-s", "--Size", type=int, help="Size of the tabu list", required=True ) # Pass the arguments to main method main(parser.parse_args())
365
"""simple docstring""" import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py UpperCAmelCase ="\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Automatic Evaluation of Machine Translation},\n booktitle = {},\n year = {2002},\n pages = {311--318}\n}\n@inproceedings{lin-och-2004-orange,\n title = \"{ORANGE}: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation\",\n author = \"Lin, Chin-Yew and\n Och, Franz Josef\",\n booktitle = \"{COLING} 2004: Proceedings of the 20th International Conference on Computational Linguistics\",\n month = \"aug 23{--}aug 27\",\n year = \"2004\",\n address = \"Geneva, Switzerland\",\n publisher = \"COLING\",\n url = \"https://www.aclweb.org/anthology/C04-1072\",\n pages = \"501--507\",\n}\n" UpperCAmelCase ="\\nBLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.\nQuality is considered to be the correspondence between a machine's output and that of a human: \"the closer a machine translation is to a professional human translation,\nthe better it is\" – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and\nremains one of the most popular automated and inexpensive metrics.\n\nScores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.\nThose scores are then averaged over the whole corpus to reach an estimate of the translation's overall quality. Intelligibility or grammatical correctness\nare not taken into account[citation needed].\n\nBLEU's output is always a number between 0 and 1. This value indicates how similar the candidate text is to the reference texts, with values closer to 1\nrepresenting more similar texts. Few human translations will attain a score of 1, since this would indicate that the candidate is identical to one of the\nreference translations. For this reason, it is not necessary to attain a score of 1. Because there are more opportunities to match, adding additional\nreference translations will increase the BLEU score.\n" UpperCAmelCase ="\nComputes BLEU score of translated segments against one or more references.\nArgs:\n predictions: list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n max_order: Maximum n-gram order to use when computing BLEU score.\n smooth: Whether or not to apply Lin et al. 2004 smoothing.\nReturns:\n 'bleu': bleu score,\n 'precisions': geometric mean of n-gram precisions,\n 'brevity_penalty': brevity penalty,\n 'length_ratio': ratio of lengths,\n 'translation_length': translation_length,\n 'reference_length': reference_length\nExamples:\n\n >>> predictions = [\n ... [\"hello\", \"there\", \"general\", \"kenobi\"], # tokenized prediction of the first sample\n ... [\"foo\", \"bar\", \"foobar\"] # tokenized prediction of the second sample\n ... ]\n >>> references = [\n ... [[\"hello\", \"there\", \"general\", \"kenobi\"], [\"hello\", \"there\", \"!\"]], # tokenized references for the first sample (2 references)\n ... [[\"foo\", \"bar\", \"foobar\"]] # tokenized references for the second sample (1 reference)\n ... ]\n >>> bleu = datasets.load_metric(\"bleu\")\n >>> results = bleu.compute(predictions=predictions, references=references)\n >>> print(results[\"bleu\"])\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase__ ( datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ,id="""token""" ) ,id="""sequence""" ), """references""": datasets.Sequence( datasets.Sequence(datasets.Value("""string""" ,id="""token""" ) ,id="""sequence""" ) ,id="""references""" ), } ) ,codebase_urls=["""https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py"""] ,reference_urls=[ """https://en.wikipedia.org/wiki/BLEU""", """https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213""", ] ,) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_=4 ,lowerCamelCase_=False ) -> str: A = compute_bleu( reference_corpus=lowerCamelCase_ ,translation_corpus=lowerCamelCase_ ,max_order=lowerCamelCase_ ,smooth=lowerCamelCase_ ) ((A) , (A) , (A) , (A) , (A) , (A)) = score return { "bleu": bleu, "precisions": precisions, "brevity_penalty": bp, "length_ratio": ratio, "translation_length": translation_length, "reference_length": reference_length, }
77
0
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch _UpperCamelCase = random.Random() def _lowercase ( lowercase__ , lowercase__=1.0 , lowercase__=None , lowercase__=None ): if rng is None: __lowerCAmelCase : Any = global_rng __lowerCAmelCase : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __lowercase (unittest.TestCase ): def __init__( self , A_ , A_=7 , A_=400 , A_=2000 , A_=10 , A_=160 , A_=8 , A_=0.0 , A_=4000 , A_=False , A_=True , ) ->List[str]: '''simple docstring''' __lowerCAmelCase : Optional[int] = parent __lowerCAmelCase : Dict = batch_size __lowerCAmelCase : str = min_seq_length __lowerCAmelCase : int = max_seq_length __lowerCAmelCase : Any = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : Any = padding_value __lowerCAmelCase : str = sampling_rate __lowerCAmelCase : Optional[Any] = return_attention_mask __lowerCAmelCase : Optional[Any] = do_normalize __lowerCAmelCase : Optional[Any] = feature_size __lowerCAmelCase : Optional[int] = chunk_length __lowerCAmelCase : Optional[Any] = hop_length def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__ ( self , A_=False , A_=False ) ->Optional[Any]: '''simple docstring''' def _flatten(A_ ): return list(itertools.chain(*A_ ) ) if equal_length: __lowerCAmelCase : str = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCAmelCase : Any = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : Optional[Any] = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowercase (_UpperCAmelCase , unittest.TestCase ): _UpperCamelCase = WhisperFeatureExtractor if is_speech_available() else None def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Tuple = WhisperFeatureExtractionTester(self ) def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : List[str] = feat_extract_first.save_pretrained(A_ )[0] check_json_file_has_correct_format(A_ ) __lowerCAmelCase : int = self.feature_extraction_class.from_pretrained(A_ ) __lowerCAmelCase : Dict = feat_extract_first.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_second.to_dict() __lowerCAmelCase : Union[str, Any] = feat_extract_first.mel_filters __lowerCAmelCase : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase : Union[str, Any] = os.path.join(A_ , '''feat_extract.json''' ) feat_extract_first.to_json_file(A_ ) __lowerCAmelCase : List[str] = self.feature_extraction_class.from_json_file(A_ ) __lowerCAmelCase : List[str] = feat_extract_first.to_dict() __lowerCAmelCase : Tuple = feat_extract_second.to_dict() __lowerCAmelCase : Any = feat_extract_first.mel_filters __lowerCAmelCase : List[str] = feat_extract_second.mel_filters self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size __lowerCAmelCase : Tuple = feature_extractor(A_ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __lowerCAmelCase : Dict = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase : Union[str, Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[Any] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase : Optional[int] = np.asarray(A_ ) __lowerCAmelCase : Dict = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) # Test truncation required __lowerCAmelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] __lowerCAmelCase : Dict = [np.asarray(A_ ) for speech_input in speech_inputs] __lowerCAmelCase : Tuple = [x[: feature_extractor.n_samples] for x in speech_inputs] __lowerCAmelCase : Optional[int] = [np.asarray(A_ ) for speech_input in speech_inputs_truncated] __lowerCAmelCase : Any = feature_extractor(A_ , return_tensors='''np''' ).input_features __lowerCAmelCase : List[str] = feature_extractor(A_ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1e-3 ) ) def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' import torch __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : List[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) __lowerCAmelCase : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Tuple = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowerCAmelCase : int = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCamelCase__ ( self , A_ ) ->str: '''simple docstring''' __lowerCAmelCase : Any = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __lowerCAmelCase : Union[str, Any] = ds.sort('''id''' ).select(range(A_ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = torch.tensor( [ 0.1_193, -0.0_946, -0.1_098, -0.0_196, 0.0_225, -0.0_690, -0.1_736, 0.0_951, 0.0_971, -0.0_817, -0.0_702, 0.0_162, 0.0_260, 0.0_017, -0.0_192, -0.1_678, 0.0_709, -0.1_867, -0.0_655, -0.0_274, -0.0_234, -0.1_884, -0.0_516, -0.0_554, -0.0_274, -0.1_425, -0.1_423, 0.0_837, 0.0_377, -0.0_854 ] ) # fmt: on __lowerCAmelCase : int = self._load_datasamples(1 ) __lowerCAmelCase : Any = WhisperFeatureExtractor() __lowerCAmelCase : Optional[Any] = feature_extractor(A_ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , A_ , atol=1e-4 ) ) def UpperCamelCase__ ( self ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = self._load_datasamples(1 )[0] __lowerCAmelCase : Optional[Any] = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue __lowerCAmelCase : Union[str, Any] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=A_ )[0] self.assertTrue(np.all(np.mean(A_ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ ) - 1 ) < 1e-3 ) )
275
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase_ = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
260
from sympy import diff, lambdify, symbols from sympy.functions import * # noqa: F403 def lowerCamelCase_ ( lowerCAmelCase: str , lowerCAmelCase: complex , lowerCAmelCase: str = "x" , lowerCAmelCase: float = 10**-10 , lowerCAmelCase: int = 1 , )-> complex: _snake_case : Optional[int] = symbols(lowerCAmelCase ) _snake_case : Tuple = lambdify(lowerCAmelCase , lowerCAmelCase ) _snake_case : Union[str, Any] = lambdify(lowerCAmelCase , diff(lowerCAmelCase , lowerCAmelCase ) ) _snake_case : int = starting_point while True: if diff_function(lowerCAmelCase ) != 0: _snake_case : Optional[Any] = prev_guess - multiplicity * func(lowerCAmelCase ) / diff_function( lowerCAmelCase ) else: raise ZeroDivisionError('Could not find root' ) from None # Precision is checked by comparing the difference of consecutive guesses if abs(next_guess - prev_guess ) < precision: return next_guess _snake_case : Dict = next_guess # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F"""The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}""") # Find root of polynomial # Find fourth Root of 5 print(F"""The root of x**4 - 5 = 0 is {newton_raphson("x**4 -5", 0.4 +5J)}""") # Find value of e print( """The root of log(y) - 1 = 0 is """, F"""{newton_raphson("log(y) - 1", 2, variable="y")}""", ) # Exponential Roots print( """The root of exp(x) - 1 = 0 is""", F"""{newton_raphson("exp(x) - 1", 10, precision=0.005)}""", ) # Find root of cos(x) print(F"""The root of cos(x) = 0 is {newton_raphson("cos(x)", 0)}""")
260
1
"""simple docstring""" from importlib import import_module from .logging import get_logger SCREAMING_SNAKE_CASE_ : Optional[Any] = get_logger(__name__) class a : """simple docstring""" def __init__( self: Union[str, Any] , UpperCamelCase: Dict , UpperCamelCase: Dict=None ): """simple docstring""" A__ = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("""__""" ): setattr(self , lowercase__ , getattr(lowercase__ , lowercase__ ) ) A__ = module._original_module if isinstance(lowercase__ , _PatchedModuleObj ) else module class a : """simple docstring""" UpperCAmelCase = [] def __init__( self: Optional[int] , UpperCamelCase: Optional[int] , UpperCamelCase: str , UpperCamelCase: Dict , UpperCamelCase: Optional[int]=None ): """simple docstring""" A__ = obj A__ = target A__ = new A__ = target.split(""".""" )[0] A__ = {} A__ = attrs or [] def __enter__( self: Tuple ): """simple docstring""" *A__ , A__ = self.target.split(""".""" ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(lowercase__ ) ): try: A__ = import_module(""".""".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): A__ = getattr(self.obj , lowercase__ ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(lowercase__ , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): A__ = obj_attr # patch at top level setattr(self.obj , lowercase__ , _PatchedModuleObj(lowercase__ , attrs=self.attrs ) ) A__ = getattr(self.obj , lowercase__ ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(lowercase__ , lowercase__ , _PatchedModuleObj(getattr(lowercase__ , lowercase__ , lowercase__ ) , attrs=self.attrs ) ) A__ = getattr(lowercase__ , lowercase__ ) # finally set the target attribute setattr(lowercase__ , lowercase__ , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: A__ = getattr(import_module(""".""".join(lowercase__ ) ) , lowercase__ ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , lowercase__ ) is attr_value: A__ = getattr(self.obj , lowercase__ ) setattr(self.obj , lowercase__ , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" A__ = globals()["""__builtins__"""][target_attr] setattr(self.obj , lowercase__ , self.new ) else: raise RuntimeError(f"""Tried to patch attribute {target_attr} instead of a submodule.""" ) def __exit__( self: Optional[Any] , *UpperCamelCase: int ): """simple docstring""" for attr in list(self.original ): setattr(self.obj , lowercase__ , self.original.pop(lowercase__ ) ) def UpperCamelCase ( self: Dict ): """simple docstring""" self.__enter__() self._active_patches.append(self ) def UpperCamelCase ( self: List[Any] ): """simple docstring""" try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
335
'''simple docstring''' import argparse import json import logging import os import shutil import sys import tempfile import unittest from unittest import mock import torch from accelerate.utils import write_basic_config from transformers.testing_utils import TestCasePlus, get_gpu_count, run_command, slow, torch_device from transformers.utils import is_apex_available logging.basicConfig(level=logging.DEBUG) lowerCAmelCase__ = logging.getLogger() def _A ( ): """simple docstring""" __lowercase = argparse.ArgumentParser() parser.add_argument('''-f''' ) __lowercase = parser.parse_args() return args.f def _A ( A__ ): """simple docstring""" __lowercase = {} __lowercase = os.path.join(A__ , '''all_results.json''' ) if os.path.exists(A__ ): with open(A__ , '''r''' ) as f: __lowercase = json.load(A__ ) else: raise ValueError(F"can't find {path}" ) return results def _A ( ): """simple docstring""" __lowercase = torch.cuda.is_available() and torch_device == '''cuda''' return is_using_cuda and is_apex_available() lowerCAmelCase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @classmethod def SCREAMING_SNAKE_CASE ( cls : List[str] ): # Write Accelerate config, will pick up on CPU, GPU, and multi-GPU __lowercase = tempfile.mkdtemp() __lowercase = os.path.join(cls.tmpdir ,'''default_config.yml''' ) write_basic_config(save_location=cls.configPath ) __lowercase = ['''accelerate''', '''launch''', '''--config_file''', cls.configPath] @classmethod def SCREAMING_SNAKE_CASE ( cls : str ): shutil.rmtree(cls.tmpdir ) @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/text-classification/run_glue_no_trainer.py\n --model_name_or_path distilbert-base-uncased\n --output_dir {tmp_dir}\n --train_file ./tests/fixtures/tests_samples/MRPC/train.csv\n --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --learning_rate=1e-4\n --seed=42\n --checkpointing_steps epoch\n --with_tracking\n ".split() if is_cuda_and_apex_available(): testargs.append('''--fp16''' ) run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] ,0.7_5 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''glue_no_trainer''' ) ) ) @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/language-modeling/run_clm_no_trainer.py\n --model_name_or_path distilgpt2\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --block_size 128\n --per_device_train_batch_size 5\n --per_device_eval_batch_size 5\n --num_train_epochs 2\n --output_dir {tmp_dir}\n --checkpointing_steps epoch\n --with_tracking\n ".split() if torch.cuda.device_count() > 1: # Skipping because there are not enough batches to train the model + would need a drop_last to work. return run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertLess(result['''perplexity'''] ,1_0_0 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''clm_no_trainer''' ) ) ) @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/language-modeling/run_mlm_no_trainer.py\n --model_name_or_path distilroberta-base\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --output_dir {tmp_dir}\n --num_train_epochs=1\n --checkpointing_steps epoch\n --with_tracking\n ".split() run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertLess(result['''perplexity'''] ,4_2 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''mlm_no_trainer''' ) ) ) @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : Tuple ): # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu __lowercase = 7 if get_gpu_count() > 1 else 2 __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/token-classification/run_ner_no_trainer.py\n --model_name_or_path bert-base-uncased\n --train_file tests/fixtures/tests_samples/conll/sample.json\n --validation_file tests/fixtures/tests_samples/conll/sample.json\n --output_dir {tmp_dir}\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=2\n --num_train_epochs={epochs}\n --seed 7\n --checkpointing_steps epoch\n --with_tracking\n ".split() run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] ,0.7_5 ) self.assertLess(result['''train_loss'''] ,0.5 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''ner_no_trainer''' ) ) ) @unittest.skip(reason='''Fix me @muellerzr''' ) @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/question-answering/run_qa_no_trainer.py\n --model_name_or_path bert-base-uncased\n --version_2_with_negative\n --train_file tests/fixtures/tests_samples/SQUAD/sample.json\n --validation_file tests/fixtures/tests_samples/SQUAD/sample.json\n --output_dir {tmp_dir}\n --seed=42\n --max_train_steps=10\n --num_warmup_steps=2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --checkpointing_steps epoch\n --with_tracking\n ".split() run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) # Because we use --version_2_with_negative the testing script uses SQuAD v2 metrics. self.assertGreaterEqual(result['''eval_f1'''] ,2_8 ) self.assertGreaterEqual(result['''eval_exact'''] ,2_8 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''qa_no_trainer''' ) ) ) @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/multiple-choice/run_swag_no_trainer.py\n --model_name_or_path bert-base-uncased\n --train_file tests/fixtures/tests_samples/swag/sample.json\n --validation_file tests/fixtures/tests_samples/swag/sample.json\n --output_dir {tmp_dir}\n --max_train_steps=20\n --num_warmup_steps=2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --with_tracking\n ".split() run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] ,0.8 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''swag_no_trainer''' ) ) ) @slow @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/summarization/run_summarization_no_trainer.py\n --model_name_or_path t5-small\n --train_file tests/fixtures/tests_samples/xsum/sample.json\n --validation_file tests/fixtures/tests_samples/xsum/sample.json\n --output_dir {tmp_dir}\n --max_train_steps=50\n --num_warmup_steps=8\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --checkpointing_steps epoch\n --with_tracking\n ".split() run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertGreaterEqual(result['''eval_rouge1'''] ,1_0 ) self.assertGreaterEqual(result['''eval_rouge2'''] ,2 ) self.assertGreaterEqual(result['''eval_rougeL'''] ,7 ) self.assertGreaterEqual(result['''eval_rougeLsum'''] ,7 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''summarization_no_trainer''' ) ) ) @slow @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/translation/run_translation_no_trainer.py\n --model_name_or_path sshleifer/student_marian_en_ro_6_1\n --source_lang en\n --target_lang ro\n --train_file tests/fixtures/tests_samples/wmt16/sample.json\n --validation_file tests/fixtures/tests_samples/wmt16/sample.json\n --output_dir {tmp_dir}\n --max_train_steps=50\n --num_warmup_steps=8\n --num_beams=6\n --learning_rate=3e-3\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --source_lang en_XX\n --target_lang ro_RO\n --checkpointing_steps epoch\n --with_tracking\n ".split() run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertGreaterEqual(result['''eval_bleu'''] ,3_0 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''translation_no_trainer''' ) ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = logging.StreamHandler(sys.stdout ) logger.addHandler(lowercase__ ) __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/semantic-segmentation/run_semantic_segmentation_no_trainer.py\n --dataset_name huggingface/semantic-segmentation-test-sample\n --output_dir {tmp_dir}\n --max_train_steps=10\n --num_warmup_steps=2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --checkpointing_steps epoch\n ".split() run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) self.assertGreaterEqual(result['''eval_overall_accuracy'''] ,0.1_0 ) @mock.patch.dict(os.environ ,{'''WANDB_MODE''': '''offline'''} ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.get_auto_remove_tmp_dir() __lowercase = F"\n {self.examples_dir}/pytorch/image-classification/run_image_classification_no_trainer.py\n --model_name_or_path google/vit-base-patch16-224-in21k\n --dataset_name hf-internal-testing/cats_vs_dogs_sample\n --learning_rate 1e-4\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 1\n --max_train_steps 2\n --train_val_split 0.1\n --seed 42\n --output_dir {tmp_dir}\n --with_tracking\n --checkpointing_steps 1\n ".split() if is_cuda_and_apex_available(): testargs.append('''--fp16''' ) run_command(self._launch_args + testargs ) __lowercase = get_results(lowercase__ ) # The base model scores a 25% self.assertGreaterEqual(result['''eval_accuracy'''] ,0.6 ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''step_1''' ) ) ) self.assertTrue(os.path.exists(os.path.join(lowercase__ ,'''image_classification_no_trainer''' ) ) )
104
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging lowerCAmelCase_ = logging.get_logger(__name__) class __lowerCAmelCase ( _a ): lowerCamelCase_ : Any = ['''pixel_values'''] def __init__(self , __magic_name__ = True , __magic_name__ = None , __magic_name__ = PILImageResampling.BILINEAR , __magic_name__ = True , __magic_name__ = None , __magic_name__ = True , __magic_name__ = 1 / 255 , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , **__magic_name__ , ) -> None: '''simple docstring''' super().__init__(**__magic_name__ ) snake_case_ : List[str] = size if size is not None else {'''shortest_edge''': 256} snake_case_ : List[Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) snake_case_ : int = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} snake_case_ : str = get_size_dict(__magic_name__ ) snake_case_ : Any = do_resize snake_case_ : List[str] = size snake_case_ : Dict = resample snake_case_ : Union[str, Any] = do_center_crop snake_case_ : Any = crop_size snake_case_ : Union[str, Any] = do_rescale snake_case_ : Tuple = rescale_factor snake_case_ : Union[str, Any] = do_normalize snake_case_ : str = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ : str = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ = PILImageResampling.BICUBIC , __magic_name__ = None , **__magic_name__ , ) -> np.ndarray: '''simple docstring''' snake_case_ : Any = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) snake_case_ : Optional[Any] = get_resize_output_image_size(__magic_name__ , size=size['''shortest_edge'''] , default_to_square=__magic_name__ ) return resize(__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ) -> np.ndarray: '''simple docstring''' snake_case_ : str = get_size_dict(__magic_name__ ) return center_crop(__magic_name__ , size=(size['''height'''], size['''width''']) , data_format=__magic_name__ , **__magic_name__ ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ ) -> np.ndarray: '''simple docstring''' return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ) -> np.ndarray: '''simple docstring''' return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def lowerCamelCase (self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = ChannelDimension.FIRST , **__magic_name__ , ) -> int: '''simple docstring''' snake_case_ : List[Any] = do_resize if do_resize is not None else self.do_resize snake_case_ : List[Any] = size if size is not None else self.size snake_case_ : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) snake_case_ : Optional[Any] = resample if resample is not None else self.resample snake_case_ : Union[str, Any] = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ : List[str] = crop_size if crop_size is not None else self.crop_size snake_case_ : Optional[int] = get_size_dict(__magic_name__ ) snake_case_ : int = do_rescale if do_rescale is not None else self.do_rescale snake_case_ : Tuple = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ : Union[str, Any] = do_normalize if do_normalize is not None else self.do_normalize snake_case_ : int = image_mean if image_mean is not None else self.image_mean snake_case_ : Dict = image_std if image_std is not None else self.image_std snake_case_ : str = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. snake_case_ : List[str] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: snake_case_ : List[Any] = [self.resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ ) for image in images] if do_center_crop: snake_case_ : Optional[int] = [self.center_crop(image=__magic_name__ , size=__magic_name__ ) for image in images] if do_rescale: snake_case_ : int = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: snake_case_ : Any = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] snake_case_ : int = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] snake_case_ : Tuple = {'''pixel_values''': images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
279
import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel lowerCAmelCase_ = logging.getLogger(__name__) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase ) -> int: """simple docstring""" if os.path.exists(_UpperCamelCase ): if os.path.exists(os.path.join(_UpperCamelCase , '''config.json''' ) ) and os.path.isfile( os.path.join(_UpperCamelCase , '''config.json''' ) ): os.remove(os.path.join(_UpperCamelCase , '''config.json''' ) ) if os.path.exists(os.path.join(_UpperCamelCase , '''pytorch_model.bin''' ) ) and os.path.isfile( os.path.join(_UpperCamelCase , '''pytorch_model.bin''' ) ): os.remove(os.path.join(_UpperCamelCase , '''pytorch_model.bin''' ) ) else: os.makedirs(_UpperCamelCase ) model.save_pretrained(_UpperCamelCase ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase=False ) -> Optional[int]: """simple docstring""" snake_case_ : List[Any] = 2 if unlogit: snake_case_ : Any = torch.pow(_UpperCamelCase , _UpperCamelCase ) snake_case_ : Optional[Any] = p * torch.log(_UpperCamelCase ) snake_case_ : Dict = 0 return -plogp.sum(dim=-1 ) def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" logger.info('''lv, h >\t''' + '''\t'''.join(f'''{x + 1}''' for x in range(len(_UpperCamelCase ) ) ) ) for row in range(len(_UpperCamelCase ) ): if tensor.dtype != torch.long: logger.info(f'''layer {row + 1}:\t''' + '''\t'''.join(f'''{x:.5f}''' for x in tensor[row].cpu().data ) ) else: logger.info(f'''layer {row + 1}:\t''' + '''\t'''.join(f'''{x:d}''' for x in tensor[row].cpu().data ) ) def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=None , _UpperCamelCase=False ) -> Union[str, Any]: """simple docstring""" snake_case_ , snake_case_ : int = model.config.num_hidden_layers, model.config.num_attention_heads snake_case_ : int = torch.zeros(_UpperCamelCase , _UpperCamelCase ).to(args.device ) snake_case_ : Optional[int] = torch.zeros(_UpperCamelCase , _UpperCamelCase ).to(args.device ) if head_mask is None: snake_case_ : Tuple = torch.ones(_UpperCamelCase , _UpperCamelCase ).to(args.device ) head_mask.requires_grad_(requires_grad=_UpperCamelCase ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: snake_case_ : Dict = None snake_case_ : Tuple = 0.0 snake_case_ : Dict = 0.0 for step, inputs in enumerate(tqdm(_UpperCamelCase , desc='''Iteration''' , disable=args.local_rank not in [-1, 0] ) ): snake_case_ : Any = tuple(t.to(args.device ) for t in inputs ) ((snake_case_) , ) : Union[str, Any] = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) snake_case_ : List[str] = model(_UpperCamelCase , labels=_UpperCamelCase , head_mask=_UpperCamelCase ) # (loss), lm_logits, presents, (all hidden_states), (attentions) snake_case_ , snake_case_ , snake_case_ : int = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(_UpperCamelCase ): snake_case_ : Dict = entropy(attn.detach() , _UpperCamelCase ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(_UpperCamelCase ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: snake_case_ : Union[str, Any] = 2 snake_case_ : Any = torch.pow(torch.pow(_UpperCamelCase , _UpperCamelCase ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1E-20 if not args.dont_normalize_global_importance: snake_case_ : Union[str, Any] = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info('''Attention entropies''' ) print_ad_tensor(_UpperCamelCase ) if compute_importance: logger.info('''Head importance scores''' ) print_ad_tensor(_UpperCamelCase ) logger.info('''Head ranked by importance scores''' ) snake_case_ : Optional[int] = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) snake_case_ : Union[str, Any] = torch.arange( head_importance.numel() , device=args.device ) snake_case_ : Dict = head_ranks.view_as(_UpperCamelCase ) print_ad_tensor(_UpperCamelCase ) return attn_entropy, head_importance, total_loss def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[int]: """simple docstring""" snake_case_ , snake_case_ , snake_case_ : Optional[int] = compute_heads_importance(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , compute_entropy=_UpperCamelCase ) snake_case_ : Any = 1 / loss # instead of downsteam score use the LM loss logger.info('''Pruning: original score: %f, threshold: %f''' , _UpperCamelCase , original_score * args.masking_threshold ) snake_case_ : Any = torch.ones_like(_UpperCamelCase ) snake_case_ : Union[str, Any] = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) snake_case_ : List[Any] = original_score while current_score >= original_score * args.masking_threshold: snake_case_ : List[str] = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads snake_case_ : Optional[Any] = float('''Inf''' ) snake_case_ : List[Any] = head_importance.view(-1 ).sort()[1] if len(_UpperCamelCase ) <= num_to_mask: print('''BREAK BY num_to_mask''' ) break # mask heads snake_case_ : Optional[int] = current_heads_to_mask[:num_to_mask] logger.info('''Heads to mask: %s''' , str(current_heads_to_mask.tolist() ) ) snake_case_ : Optional[Any] = new_head_mask.view(-1 ) snake_case_ : int = 0.0 snake_case_ : List[Any] = new_head_mask.view_as(_UpperCamelCase ) snake_case_ : List[str] = new_head_mask.clone().detach() print_ad_tensor(_UpperCamelCase ) # Compute metric and head importance again snake_case_ , snake_case_ , snake_case_ : str = compute_heads_importance( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , compute_entropy=_UpperCamelCase , head_mask=_UpperCamelCase ) snake_case_ : Tuple = 1 / loss logger.info( '''Masking: current score: %f, remaining heads %d (%.1f percents)''' , _UpperCamelCase , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info('''Final head mask''' ) print_ad_tensor(_UpperCamelCase ) np.save(os.path.join(args.output_dir , '''head_mask.npy''' ) , head_mask.detach().cpu().numpy() ) return head_mask def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> str: """simple docstring""" snake_case_ : str = datetime.now() snake_case_ , snake_case_ , snake_case_ : List[Any] = compute_heads_importance( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , compute_entropy=_UpperCamelCase , compute_importance=_UpperCamelCase , head_mask=_UpperCamelCase ) snake_case_ : Union[str, Any] = 1 / loss snake_case_ : Union[str, Any] = datetime.now() - before_time snake_case_ : int = sum(p.numel() for p in model.parameters() ) snake_case_ : Tuple = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(_UpperCamelCase ) ) } for k, v in heads_to_prune.items(): if isinstance(_UpperCamelCase , _UpperCamelCase ): snake_case_ : Any = [ v, ] assert sum(len(_UpperCamelCase ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(_UpperCamelCase ) snake_case_ : Union[str, Any] = sum(p.numel() for p in model.parameters() ) snake_case_ : Dict = datetime.now() snake_case_ , snake_case_ , snake_case_ : Union[str, Any] = compute_heads_importance( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , compute_entropy=_UpperCamelCase , compute_importance=_UpperCamelCase , head_mask=_UpperCamelCase , actually_pruned=_UpperCamelCase , ) snake_case_ : Union[str, Any] = 1 / loss snake_case_ : Optional[Any] = datetime.now() - before_time logger.info( '''Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)''' , _UpperCamelCase , _UpperCamelCase , pruned_num_params / original_num_params * 100 , ) logger.info('''Pruning: score with masking: %f score with pruning: %f''' , _UpperCamelCase , _UpperCamelCase ) logger.info('''Pruning: speed ratio (original timing / new timing): %f percents''' , original_time / new_time * 100 ) save_model(_UpperCamelCase , args.output_dir ) def lowerCamelCase_ ( ) -> Optional[int]: """simple docstring""" snake_case_ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--data_dir''' , default=_UpperCamelCase , type=_UpperCamelCase , required=_UpperCamelCase , help='''The input data dir. Should contain the .tsv files (or other data files) for the task.''' , ) parser.add_argument( '''--model_name_or_path''' , default=_UpperCamelCase , type=_UpperCamelCase , required=_UpperCamelCase , help='''Path to pretrained model or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--output_dir''' , default=_UpperCamelCase , type=_UpperCamelCase , required=_UpperCamelCase , help='''The output directory where the model predictions and checkpoints will be written.''' , ) # Other parameters parser.add_argument( '''--config_name''' , default='''''' , type=_UpperCamelCase , help='''Pretrained config name or path if not the same as model_name_or_path''' , ) parser.add_argument( '''--tokenizer_name''' , default='''''' , type=_UpperCamelCase , help='''Pretrained tokenizer name or path if not the same as model_name_or_path''' , ) parser.add_argument( '''--cache_dir''' , default=_UpperCamelCase , type=_UpperCamelCase , help='''Where do you want to store the pre-trained models downloaded from s3''' , ) parser.add_argument( '''--data_subset''' , type=_UpperCamelCase , default=-1 , help='''If > 0: limit the data to a subset of data_subset instances.''' ) parser.add_argument( '''--overwrite_output_dir''' , action='''store_true''' , help='''Whether to overwrite data in output directory''' ) parser.add_argument( '''--overwrite_cache''' , action='''store_true''' , help='''Overwrite the cached training and evaluation sets''' ) parser.add_argument( '''--dont_normalize_importance_by_layer''' , action='''store_true''' , help='''Don\'t normalize importance score by layers''' ) parser.add_argument( '''--dont_normalize_global_importance''' , action='''store_true''' , help='''Don\'t normalize all importance scores between 0 and 1''' , ) parser.add_argument( '''--try_masking''' , action='''store_true''' , help='''Whether to try to mask head until a threshold of accuracy.''' ) parser.add_argument( '''--masking_threshold''' , default=0.9 , type=_UpperCamelCase , help='''masking threshold in term of metrics (stop masking when metric < threshold * original metric value).''' , ) parser.add_argument( '''--masking_amount''' , default=0.1 , type=_UpperCamelCase , help='''Amount to heads to masking at each masking step.''' ) parser.add_argument('''--metric_name''' , default='''acc''' , type=_UpperCamelCase , help='''Metric to use for head masking.''' ) parser.add_argument( '''--max_seq_length''' , default=128 , type=_UpperCamelCase , help=( '''The maximum total input sequence length after WordPiece tokenization. \n''' '''Sequences longer than this will be truncated, sequences shorter padded.''' ) , ) parser.add_argument('''--batch_size''' , default=1 , type=_UpperCamelCase , help='''Batch size.''' ) parser.add_argument('''--seed''' , type=_UpperCamelCase , default=42 ) parser.add_argument('''--local_rank''' , type=_UpperCamelCase , default=-1 , help='''local_rank for distributed training on gpus''' ) parser.add_argument('''--no_cuda''' , action='''store_true''' , help='''Whether not to use CUDA when available''' ) parser.add_argument('''--server_ip''' , type=_UpperCamelCase , default='''''' , help='''Can be used for distant debugging.''' ) parser.add_argument('''--server_port''' , type=_UpperCamelCase , default='''''' , help='''Can be used for distant debugging.''' ) snake_case_ : Any = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('''Waiting for debugger attach''' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_UpperCamelCase ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: snake_case_ : Tuple = torch.device('''cuda''' if torch.cuda.is_available() and not args.no_cuda else '''cpu''' ) snake_case_ : Tuple = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) snake_case_ : List[str] = torch.device('''cuda''' , args.local_rank ) snake_case_ : Union[str, Any] = 1 torch.distributed.init_process_group(backend='''nccl''' ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info('''device: {} n_gpu: {}, distributed: {}'''.format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) snake_case_ : int = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: snake_case_ : Any = nn.parallel.DistributedDataParallel( _UpperCamelCase , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=_UpperCamelCase ) elif args.n_gpu > 1: snake_case_ : Dict = nn.DataParallel(_UpperCamelCase ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=_UpperCamelCase ) torch.save(_UpperCamelCase , os.path.join(args.output_dir , '''run_args.bin''' ) ) logger.info('''Training/evaluation parameters %s''' , _UpperCamelCase ) # Prepare dataset snake_case_ : str = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) snake_case_ : Any = (torch.from_numpy(_UpperCamelCase ),) snake_case_ : Any = TensorDataset(*_UpperCamelCase ) snake_case_ : List[str] = RandomSampler(_UpperCamelCase ) snake_case_ : int = DataLoader(_UpperCamelCase , sampler=_UpperCamelCase , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: snake_case_ : List[str] = mask_heads(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) prune_heads(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) if __name__ == "__main__": main()
279
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A ={ 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A =['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A =[ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A =[ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A =[ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys A =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
34
'''simple docstring''' from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class _a ( __a ): __a : str = ["""vqvae"""] def __init__( self : str , lowercase : AutoencoderKL , lowercase : UNetaDConditionModel , lowercase : Mel , lowercase : Union[DDIMScheduler, DDPMScheduler] , ): '''simple docstring''' super().__init__() self.register_modules(unet=lowercase , scheduler=lowercase , mel=lowercase , vqvae=lowercase ) def A ( self : Optional[Any] ): '''simple docstring''' return 50 if isinstance(self.scheduler , lowercase ) else 1_000 @torch.no_grad() def __call__( self : Optional[Any] , lowercase : int = 1 , lowercase : str = None , lowercase : np.ndarray = None , lowercase : int = 0 , lowercase : int = 0 , lowercase : int = None , lowercase : torch.Generator = None , lowercase : float = 0 , lowercase : float = 0 , lowercase : torch.Generator = None , lowercase : float = 0 , lowercase : torch.Tensor = None , lowercase : torch.Tensor = None , lowercase : Tuple=True , ): '''simple docstring''' UpperCAmelCase = steps or self.get_default_steps() self.scheduler.set_timesteps(lowercase ) UpperCAmelCase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: UpperCAmelCase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: UpperCAmelCase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=lowercase , device=self.device , ) UpperCAmelCase = noise UpperCAmelCase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(lowercase , lowercase ) UpperCAmelCase = self.mel.audio_slice_to_image(lowercase ) UpperCAmelCase = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) UpperCAmelCase = (input_image / 255) * 2 - 1 UpperCAmelCase = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: UpperCAmelCase = self.vqvae.encode(torch.unsqueeze(lowercase , 0 ) ).latent_dist.sample( generator=lowercase )[0] UpperCAmelCase = self.vqvae.config.scaling_factor * input_images if start_step > 0: UpperCAmelCase = self.scheduler.add_noise(lowercase , lowercase , self.scheduler.timesteps[start_step - 1] ) UpperCAmelCase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) UpperCAmelCase = int(mask_start_secs * pixels_per_second ) UpperCAmelCase = int(mask_end_secs * pixels_per_second ) UpperCAmelCase = self.scheduler.add_noise(lowercase , lowercase , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , lowercase ): UpperCAmelCase = self.unet(lowercase , lowercase , lowercase )['''sample'''] else: UpperCAmelCase = self.unet(lowercase , lowercase )['''sample'''] if isinstance(self.scheduler , lowercase ): UpperCAmelCase = self.scheduler.step( model_output=lowercase , timestep=lowercase , sample=lowercase , eta=lowercase , generator=lowercase , )['''prev_sample'''] else: UpperCAmelCase = self.scheduler.step( model_output=lowercase , timestep=lowercase , sample=lowercase , generator=lowercase , )['''prev_sample'''] if mask is not None: if mask_start > 0: UpperCAmelCase = mask[:, step, :, :mask_start] if mask_end > 0: UpperCAmelCase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance UpperCAmelCase = 1 / self.vqvae.config.scaling_factor * images UpperCAmelCase = self.vqvae.decode(lowercase )['''sample'''] UpperCAmelCase = (images / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() UpperCAmelCase = (images * 255).round().astype('''uint8''' ) UpperCAmelCase = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(lowercase , mode='''RGB''' ).convert('''L''' ) for _ in images) ) UpperCAmelCase = [self.mel.image_to_audio(lowercase ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(lowercase )[:, np.newaxis, :] ) , **ImagePipelineOutput(lowercase ) ) @torch.no_grad() def A ( self : Dict , lowercase : List[Image.Image] , lowercase : int = 50 ): '''simple docstring''' assert isinstance(self.scheduler , lowercase ) self.scheduler.set_timesteps(lowercase ) UpperCAmelCase = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) UpperCAmelCase = (sample / 255) * 2 - 1 UpperCAmelCase = torch.Tensor(lowercase ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): UpperCAmelCase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps UpperCAmelCase = self.scheduler.alphas_cumprod[t] UpperCAmelCase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) UpperCAmelCase = 1 - alpha_prod_t UpperCAmelCase = self.unet(lowercase , lowercase )['''sample'''] UpperCAmelCase = (1 - alpha_prod_t_prev) ** 0.5 * model_output UpperCAmelCase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) UpperCAmelCase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def A ( lowercase : torch.Tensor , lowercase : torch.Tensor , lowercase : float ): '''simple docstring''' UpperCAmelCase = acos(torch.dot(torch.flatten(lowercase ) , torch.flatten(lowercase ) ) / torch.norm(lowercase ) / torch.norm(lowercase ) ) return sin((1 - alpha) * theta ) * xa / sin(lowercase ) + sin(alpha * theta ) * xa / sin(lowercase )
34
1
"""simple docstring""" from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer lowerCamelCase_ : Any = logging.get_logger(__name__) lowerCamelCase_ : int = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } lowerCamelCase_ : Any = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } lowerCamelCase_ : List[Any] = { """facebook/blenderbot_small-90M""": 5_12, } class _UpperCamelCase ( _A ): '''simple docstring''' __UpperCamelCase : Union[str, Any] = VOCAB_FILES_NAMES __UpperCamelCase : Tuple = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : int = BlenderbotSmallTokenizer def __init__( self : Union[str, Any] , snake_case_ : Union[str, Any]=None , snake_case_ : int=None , snake_case_ : str="<|endoftext|>" , snake_case_ : Optional[Any]="<|endoftext|>" , snake_case_ : Any="<|endoftext|>" , snake_case_ : List[Any]=False , snake_case_ : int=True , **snake_case_ : Dict , ): super().__init__( ByteLevelBPETokenizer( vocab=snake_case_ , merges=snake_case_ , add_prefix_space=snake_case_ , trim_offsets=snake_case_ , ) , bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , **snake_case_ , ) UpperCamelCase_: Optional[Any] = add_prefix_space def lowerCAmelCase__ ( self : Optional[Any] , snake_case_ : Dict , snake_case_ : List[Any]=None ): UpperCamelCase_: Any = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowerCAmelCase__ ( self : Any , snake_case_ : List[int] , snake_case_ : Optional[List[int]] = None ): UpperCamelCase_: Tuple = [self.sep_token_id] UpperCamelCase_: Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
367
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration lowerCamelCase_ : Any = 5_00_00 lowerCamelCase_ : Any = 50_00 lowerCamelCase_ , lowerCamelCase_ : Tuple = os.path.split(__file__) lowerCamelCase_ : int = os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def A__ ( lowerCamelCase , lowerCamelCase ) -> Tuple: for i in range(lowerCamelCase ): UpperCamelCase_: Dict = dataset[i] @get_duration def A__ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> Tuple: for i in range(0 , len(lowerCamelCase ) , lowerCamelCase ): UpperCamelCase_: List[Any] = dataset[i : i + batch_size] @get_duration def A__ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> Any: with dataset.formatted_as(type=lowerCamelCase ): for i in range(lowerCamelCase ): UpperCamelCase_: List[str] = dataset[i] @get_duration def A__ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> Optional[int]: with dataset.formatted_as(type=lowerCamelCase ): for i in range(0 , lowerCamelCase , lowerCamelCase ): UpperCamelCase_: Union[str, Any] = dataset[i : i + batch_size] def A__ ( ) -> Tuple: UpperCamelCase_: int = {"""num examples""": SPEED_TEST_N_EXAMPLES} UpperCamelCase_: Union[str, Any] = [ (read, {"""length""": SMALL_TEST}), (read, {"""length""": SPEED_TEST_N_EXAMPLES}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 1_00}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10_00}), (read_formatted, {"""type""": """numpy""", """length""": SMALL_TEST}), (read_formatted, {"""type""": """pandas""", """length""": SMALL_TEST}), (read_formatted, {"""type""": """torch""", """length""": SMALL_TEST}), (read_formatted, {"""type""": """tensorflow""", """length""": SMALL_TEST}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10_00}), ] UpperCamelCase_: Tuple = [ (read, {"""length""": SMALL_TEST}), (read, {"""length""": SPEED_TEST_N_EXAMPLES}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 1_00}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10_00}), (read_formatted, {"""type""": """numpy""", """length""": SMALL_TEST}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10_00}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("""generating dataset""" ) UpperCamelCase_: int = datasets.Features( {"""list""": datasets.Sequence(datasets.Value("""float32""" ) ), """numbers""": datasets.Value("""float32""" )} ) UpperCamelCase_: Optional[int] = generate_example_dataset( os.path.join(lowerCamelCase , """dataset.arrow""" ) , lowerCamelCase , num_examples=lowerCamelCase , seq_shapes={"""list""": (1_00,)} , ) print("""first set of iterations""" ) for func, kwargs in functions: print(func.__name__ , str(lowerCamelCase ) ) UpperCamelCase_: List[Any] = func(lowerCamelCase , **lowerCamelCase ) print("""shuffling dataset""" ) UpperCamelCase_: Dict = dataset.shuffle() print("""Second set of iterations (after shuffling""" ) for func, kwargs in functions_shuffled: print("""shuffled """ , func.__name__ , str(lowerCamelCase ) ) UpperCamelCase_: List[Any] = func( lowerCamelCase , **lowerCamelCase ) with open(lowerCamelCase , """wb""" ) as f: f.write(json.dumps(lowerCamelCase ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
223
0
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> Tuple: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "ZinengTang/tvlt-base" __lowerCAmelCase : Dict = tempfile.mkdtemp() def _SCREAMING_SNAKE_CASE ( self: List[str] , **_SCREAMING_SNAKE_CASE: Optional[Any]) -> Any: """simple docstring""" return TvltImageProcessor.from_pretrained(self.checkpoint , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: str , **_SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" return TvltFeatureExtractor.from_pretrained(self.checkpoint , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Dict) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname) def _SCREAMING_SNAKE_CASE ( self: Tuple) -> int: """simple docstring""" __lowerCAmelCase : Tuple = self.get_image_processor() __lowerCAmelCase : List[Any] = self.get_feature_extractor() __lowerCAmelCase : Any = TvltProcessor(image_processor=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE) processor.save_pretrained(self.tmpdirname) __lowerCAmelCase : int = TvltProcessor.from_pretrained(self.tmpdirname) self.assertIsInstance(processor.feature_extractor , _SCREAMING_SNAKE_CASE) self.assertIsInstance(processor.image_processor , _SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Optional[int]: """simple docstring""" __lowerCAmelCase : str = self.get_image_processor() __lowerCAmelCase : Tuple = self.get_feature_extractor() __lowerCAmelCase : str = TvltProcessor(image_processor=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = np.ones([1_2000]) __lowerCAmelCase : List[Any] = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : List[str] = processor(audio=_SCREAMING_SNAKE_CASE , return_tensors="np") for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> int: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_image_processor() __lowerCAmelCase : Any = self.get_feature_extractor() __lowerCAmelCase : Dict = TvltProcessor(image_processor=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[str] = np.ones([3, 224, 224]) __lowerCAmelCase : Union[str, Any] = image_processor(_SCREAMING_SNAKE_CASE , return_tensors="np") __lowerCAmelCase : Optional[int] = processor(images=_SCREAMING_SNAKE_CASE , return_tensors="np") for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2) def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> int: """simple docstring""" __lowerCAmelCase : Dict = self.get_image_processor() __lowerCAmelCase : List[str] = self.get_feature_extractor() __lowerCAmelCase : int = TvltProcessor(image_processor=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = np.ones([1_2000]) __lowerCAmelCase : Any = np.ones([3, 224, 224]) __lowerCAmelCase : int = processor(audio=_SCREAMING_SNAKE_CASE , images=_SCREAMING_SNAKE_CASE) self.assertListEqual(list(inputs.keys()) , ["audio_values", "audio_mask", "pixel_values", "pixel_mask"]) # test if it raises when no input is passed with pytest.raises(_SCREAMING_SNAKE_CASE): processor() def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[Any]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_image_processor() __lowerCAmelCase : Dict = self.get_feature_extractor() __lowerCAmelCase : int = TvltProcessor(image_processor=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg="`processor` and `image_processor`+`feature_extractor` model input names do not match" , )
269
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Any = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'mctct' def __init__( self: Union[str, Any] , _SCREAMING_SNAKE_CASE: str=8065 , _SCREAMING_SNAKE_CASE: str=1536 , _SCREAMING_SNAKE_CASE: str=36 , _SCREAMING_SNAKE_CASE: Optional[Any]=6144 , _SCREAMING_SNAKE_CASE: Optional[Any]=4 , _SCREAMING_SNAKE_CASE: Union[str, Any]=384 , _SCREAMING_SNAKE_CASE: Optional[Any]=920 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1e-5 , _SCREAMING_SNAKE_CASE: List[Any]=0.3 , _SCREAMING_SNAKE_CASE: Optional[Any]="relu" , _SCREAMING_SNAKE_CASE: Optional[int]=0.02 , _SCREAMING_SNAKE_CASE: Optional[Any]=0.3 , _SCREAMING_SNAKE_CASE: Dict=0.3 , _SCREAMING_SNAKE_CASE: List[Any]=1 , _SCREAMING_SNAKE_CASE: Optional[Any]=0 , _SCREAMING_SNAKE_CASE: List[str]=2 , _SCREAMING_SNAKE_CASE: Union[str, Any]=1 , _SCREAMING_SNAKE_CASE: Tuple=0.3 , _SCREAMING_SNAKE_CASE: Dict=1 , _SCREAMING_SNAKE_CASE: int=(7,) , _SCREAMING_SNAKE_CASE: str=(3,) , _SCREAMING_SNAKE_CASE: Union[str, Any]=80 , _SCREAMING_SNAKE_CASE: Tuple=1 , _SCREAMING_SNAKE_CASE: Dict=None , _SCREAMING_SNAKE_CASE: Tuple="sum" , _SCREAMING_SNAKE_CASE: List[str]=False , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Tuple: """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = vocab_size __lowerCAmelCase : str = hidden_size __lowerCAmelCase : str = num_hidden_layers __lowerCAmelCase : str = intermediate_size __lowerCAmelCase : List[Any] = num_attention_heads __lowerCAmelCase : Dict = attention_head_dim __lowerCAmelCase : Optional[int] = max_position_embeddings __lowerCAmelCase : str = layer_norm_eps __lowerCAmelCase : Tuple = layerdrop __lowerCAmelCase : str = hidden_act __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob __lowerCAmelCase : str = pad_token_id __lowerCAmelCase : Optional[int] = bos_token_id __lowerCAmelCase : Union[str, Any] = eos_token_id __lowerCAmelCase : Any = conv_glu_dim __lowerCAmelCase : Optional[int] = conv_dropout __lowerCAmelCase : Union[str, Any] = num_conv_layers __lowerCAmelCase : Optional[int] = input_feat_per_channel __lowerCAmelCase : Union[str, Any] = input_channels __lowerCAmelCase : Optional[Any] = conv_channels __lowerCAmelCase : Dict = ctc_loss_reduction __lowerCAmelCase : int = ctc_zero_infinity # prevents config testing fail with exporting to json __lowerCAmelCase : List[str] = list(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = list(_SCREAMING_SNAKE_CASE) if len(self.conv_kernel) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel)` == `config.num_conv_layers` " F"""but is `len(config.conv_kernel) = {len(self.conv_kernel)}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""")
269
1
from __future__ import annotations import numpy as np def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> Tuple: '''simple docstring''' return np.maximum(0, _UpperCAmelCase ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
323
from __future__ import annotations from typing import Any class __A : def __init__( self : Optional[Any] , UpperCAmelCase_ : int ): lowerCAmelCase : Tuple = num_of_nodes lowerCAmelCase : list[list[int]] = [] lowerCAmelCase : dict[int, int] = {} def lowercase__ ( self : List[str] , UpperCAmelCase_ : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): self.m_edges.append([u_node, v_node, weight] ) def lowercase__ ( self : Dict , UpperCAmelCase_ : int ): if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def lowercase__ ( self : Optional[int] , UpperCAmelCase_ : int ): if self.m_component[u_node] != u_node: for k in self.m_component: lowerCAmelCase : Dict = self.find_component(UpperCAmelCase_ ) def lowercase__ ( self : List[str] , UpperCAmelCase_ : list[int] , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): if component_size[u_node] <= component_size[v_node]: lowerCAmelCase : Optional[int] = v_node component_size[v_node] += component_size[u_node] self.set_component(UpperCAmelCase_ ) elif component_size[u_node] >= component_size[v_node]: lowerCAmelCase : Union[str, Any] = self.find_component(UpperCAmelCase_ ) component_size[u_node] += component_size[v_node] self.set_component(UpperCAmelCase_ ) def lowercase__ ( self : str ): lowerCAmelCase : str = [] lowerCAmelCase : Tuple = 0 lowerCAmelCase : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) lowerCAmelCase : int = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase : Union[str, Any] = edge lowerCAmelCase : Optional[int] = self.m_component[u] lowerCAmelCase : str = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): lowerCAmelCase : str = [u, v, w] for edge in minimum_weight_edge: if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCAmelCase , lowerCAmelCase , lowerCAmelCase : Any = edge lowerCAmelCase : Optional[Any] = self.m_component[u] lowerCAmelCase : Optional[Any] = self.m_component[v] if u_component != v_component: mst_weight += w self.union(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) print(f"Added edge [{u} - {v}]\nAdded weight: {w}\n" ) num_of_components -= 1 lowerCAmelCase : Optional[Any] = [-1] * self.m_num_of_nodes print(f"The total weight of the minimal spanning tree is: {mst_weight}" ) def SCREAMING_SNAKE_CASE__ ( ) -> None: '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
323
1
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/config.json""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/config.json""", } class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = """xlnet""" __lowerCAmelCase = ["""mems"""] __lowerCAmelCase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Any , lowerCamelCase_ : Any=3_2000 , lowerCamelCase_ : Dict=1024 , lowerCamelCase_ : List[str]=24 , lowerCamelCase_ : Dict=16 , lowerCamelCase_ : List[Any]=4096 , lowerCamelCase_ : List[Any]="gelu" , lowerCamelCase_ : List[Any]=True , lowerCamelCase_ : Union[str, Any]="bi" , lowerCamelCase_ : Optional[Any]=0.0_2 , lowerCamelCase_ : Optional[int]=1E-12 , lowerCamelCase_ : List[Any]=0.1 , lowerCamelCase_ : Union[str, Any]=512 , lowerCamelCase_ : Any=None , lowerCamelCase_ : str=True , lowerCamelCase_ : List[str]=False , lowerCamelCase_ : List[str]=False , lowerCamelCase_ : Optional[int]=-1 , lowerCamelCase_ : List[str]=False , lowerCamelCase_ : Union[str, Any]="last" , lowerCamelCase_ : List[str]=True , lowerCamelCase_ : str="tanh" , lowerCamelCase_ : Any=0.1 , lowerCamelCase_ : Dict=5 , lowerCamelCase_ : str=5 , lowerCamelCase_ : Optional[int]=5 , lowerCamelCase_ : Any=1 , lowerCamelCase_ : int=2 , **lowerCamelCase_ : List[Any] , ): """simple docstring""" UpperCamelCase = vocab_size UpperCamelCase = d_model UpperCamelCase = n_layer UpperCamelCase = n_head if d_model % n_head != 0: raise ValueError(f"""'d_model % n_head' ({d_model % n_head}) should be equal to 0""" ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( f"""`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})""" ) UpperCamelCase = d_model // n_head UpperCamelCase = ff_activation UpperCamelCase = d_inner UpperCamelCase = untie_r UpperCamelCase = attn_type UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = dropout UpperCamelCase = mem_len UpperCamelCase = reuse_len UpperCamelCase = bi_data UpperCamelCase = clamp_len UpperCamelCase = same_length UpperCamelCase = summary_type UpperCamelCase = summary_use_proj UpperCamelCase = summary_activation UpperCamelCase = summary_last_dropout UpperCamelCase = start_n_top UpperCamelCase = end_n_top UpperCamelCase = bos_token_id UpperCamelCase = pad_token_id UpperCamelCase = eos_token_id if "use_cache" in kwargs: warnings.warn( """The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`""" """ instead.""" , lowerCamelCase_ , ) UpperCamelCase = kwargs["""use_cache"""] UpperCamelCase = use_mems_eval UpperCamelCase = use_mems_train super().__init__(pad_token_id=lowerCamelCase_ , bos_token_id=lowerCamelCase_ , eos_token_id=lowerCamelCase_ , **lowerCamelCase_ ) @property def lowerCamelCase_ ( self : Dict ): """simple docstring""" logger.info(f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def lowerCamelCase_ ( self : Optional[int] , lowerCamelCase_ : Optional[int] ): """simple docstring""" raise NotImplementedError( f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
343
import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) _SCREAMING_SNAKE_CASE = models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(3_2, (3, 3), input_shape=(6_4, 6_4, 3), activation="""relu""") ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(3_2, (3, 3), activation="""relu""")) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=1_2_8, activation="""relu""")) classifier.add(layers.Dense(units=1, activation="""sigmoid""")) # Compiling the CNN classifier.compile( optimizer="""adam""", loss="""binary_crossentropy""", metrics=["""accuracy"""] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') _SCREAMING_SNAKE_CASE = tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 2_5_5, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) _SCREAMING_SNAKE_CASE = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 2_5_5) _SCREAMING_SNAKE_CASE = train_datagen.flow_from_directory( """dataset/training_set""", target_size=(6_4, 6_4), batch_size=3_2, class_mode="""binary""" ) _SCREAMING_SNAKE_CASE = test_datagen.flow_from_directory( """dataset/test_set""", target_size=(6_4, 6_4), batch_size=3_2, class_mode="""binary""" ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=3_0, validation_data=test_set ) classifier.save("""cnn.h5""") # Part 3 - Making new predictions _SCREAMING_SNAKE_CASE = tf.keras.preprocessing.image.load_img( """dataset/single_prediction/image.png""", target_size=(6_4, 6_4) ) _SCREAMING_SNAKE_CASE = tf.keras.preprocessing.image.img_to_array(test_image) _SCREAMING_SNAKE_CASE = np.expand_dims(test_image, axis=0) _SCREAMING_SNAKE_CASE = classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: _SCREAMING_SNAKE_CASE = """Normal""" if result[0][0] == 1: _SCREAMING_SNAKE_CASE = """Abnormality detected"""
343
1
UpperCamelCase__ = range(2, 20 + 1) UpperCamelCase__ = [10**k for k in range(ks[-1] + 1)] UpperCamelCase__ = {} def _UpperCamelCase (a__ :Optional[Any] , a__ :Union[str, Any] , a__ :Any , a__ :Optional[int] ): """simple docstring""" UpperCamelCase__ = sum(a_i[j] for j in range(a__ , len(a__ ) ) ) UpperCamelCase__ = sum(a_i[j] * base[j] for j in range(min(len(a__ ) , a__ ) ) ) UpperCamelCase__ , UpperCamelCase__ = 0, 0 UpperCamelCase__ = n - i UpperCamelCase__ = memo.get(a__ ) if sub_memo is not None: UpperCamelCase__ = sub_memo.get(a__ ) if jumps is not None and len(a__ ) > 0: # find and make the largest jump without going over UpperCamelCase__ = -1 for _k in range(len(a__ ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: UpperCamelCase__ = _k break if max_jump >= 0: UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = jumps[max_jump] # since the difference between jumps is cached, add c UpperCamelCase__ = diff + c for j in range(min(a__ , len(a__ ) ) ): UpperCamelCase__ , UpperCamelCase__ = divmod(a__ , 10 ) if new_c > 0: add(a__ , a__ , a__ ) else: UpperCamelCase__ = [] else: UpperCamelCase__ = {c: []} UpperCamelCase__ = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps UpperCamelCase__ , UpperCamelCase__ = next_term(a__ , k - 1 , i + dn , a__ ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead UpperCamelCase__ , UpperCamelCase__ = compute(a__ , a__ , i + dn , a__ ) diff += _diff dn += terms_jumped UpperCamelCase__ = sub_memo[c] # keep jumps sorted by # of terms skipped UpperCamelCase__ = 0 while j < len(a__ ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(a__ , (diff, dn, k) ) return (diff, dn) def _UpperCamelCase (a__ :Any , a__ :str , a__ :Optional[int] , a__ :List[Any] ): """simple docstring""" if i >= n: return 0, i if k > len(a__ ): a_i.extend([0 for _ in range(k - len(a__ ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) UpperCamelCase__ = i UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = 0, 0, 0 for j in range(len(a__ ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 UpperCamelCase__ = ds_c + ds_b diff += addend UpperCamelCase__ = 0 for j in range(a__ ): UpperCamelCase__ = a_i[j] + addend UpperCamelCase__ , UpperCamelCase__ = divmod(a__ , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(a__ , a__ , a__ ) return diff, i - start_i def _UpperCamelCase (a__ :Optional[int] , a__ :Dict , a__ :Dict ): """simple docstring""" for j in range(a__ , len(a__ ) ): UpperCamelCase__ = digits[j] + addend if s >= 10: UpperCamelCase__ , UpperCamelCase__ = divmod(a__ , 10 ) UpperCamelCase__ = addend // 10 + quotient else: UpperCamelCase__ = s UpperCamelCase__ = addend // 10 if addend == 0: break while addend > 0: UpperCamelCase__ , UpperCamelCase__ = divmod(a__ , 10 ) digits.append(a__ ) def _UpperCamelCase (a__ :int = 10**15 ): """simple docstring""" UpperCamelCase__ = [1] UpperCamelCase__ = 1 UpperCamelCase__ = 0 while True: UpperCamelCase__ , UpperCamelCase__ = next_term(a__ , 20 , i + dn , a__ ) dn += terms_jumped if dn == n - i: break UpperCamelCase__ = 0 for j in range(len(a__ ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(f"""{solution() = }""")
87
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy UpperCamelCase__ = logging.getLogger(__name__) UpperCamelCase__ = "pytorch_model.bin" @dataclasses.dataclass class __SCREAMING_SNAKE_CASE : snake_case : str = dataclasses.field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models."""} ) snake_case : Optional[str] = dataclasses.field( default=_a , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co."""} , ) @dataclasses.dataclass class __SCREAMING_SNAKE_CASE : snake_case : str = dataclasses.field(metadata={"""help""": """A csv or a json file containing the training data."""} ) snake_case : str = dataclasses.field(metadata={"""help""": """A csv or a json file containing the data to predict on."""} ) snake_case : Optional[str] = dataclasses.field( default=_a , metadata={"""help""": """A csv or a json file containing the validation data."""} ) snake_case : Optional[str] = dataclasses.field( default=_a , metadata={"""help""": """The name of the task to train on."""} , ) snake_case : Optional[List[str]] = dataclasses.field( default=_a , metadata={"""help""": """The list of labels for the task."""} ) @dataclasses.dataclass class __SCREAMING_SNAKE_CASE : snake_case : str = dataclasses.field( metadata={"""help""": """The output directory where the model predictions and checkpoints will be written."""} ) snake_case : Optional[str] = dataclasses.field( default="""accuracy""" , metadata={"""help""": """The evaluation metric used for the task."""} ) snake_case : Optional[str] = dataclasses.field( default="""no""" , metadata={ """help""": """The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]""" } , ) snake_case : Optional[int] = dataclasses.field( default=10 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) snake_case : Optional[float] = dataclasses.field( default=0.0 , metadata={ """help""": """How much the specified evaluation metric must improve to satisfy early stopping conditions.""" } , ) snake_case : Optional[bool] = dataclasses.field( default=_a , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the confidence score."""} , ) snake_case : Optional[bool] = dataclasses.field( default=_a , metadata={"""help""": """Whether to filter the pseudo-labeled data based on the validation performance."""} , ) snake_case : Optional[bool] = dataclasses.field( default=_a , metadata={"""help""": """Whether to fine-tune on labeled data after pseudo training."""} , ) snake_case : Optional[float] = dataclasses.field( default=0.0 , metadata={"""help""": """Confidence threshold for pseudo-labeled data filtering."""} , ) snake_case : Optional[int] = dataclasses.field( default=1_00 , metadata={"""help""": """Number of evaluation calls with no improvement after which training will be stopped."""} , ) snake_case : Optional[int] = dataclasses.field( default=_a , metadata={"""help""": """Random seed for initialization."""} , ) def _UpperCamelCase (a__ :List[str] , a__ :str , a__ :Any , a__ :Any , a__ :List[str] , a__ :Union[str, Any] ): """simple docstring""" UpperCamelCase__ = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: UpperCamelCase__ = dataset.filter(lambda a__ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 UpperCamelCase__ = int(eval_result * len(a__ ) ) print(a__ ) UpperCamelCase__ = dataset.sort("""probability""" , reverse=a__ ) UpperCamelCase__ = dataset.select(range(a__ ) ) UpperCamelCase__ = dataset.remove_columns(["""label""", """probability"""] ) UpperCamelCase__ = dataset.rename_column("""prediction""" , """label""" ) UpperCamelCase__ = dataset.map(lambda a__ : {"label": idalabel[example["label"]]} ) UpperCamelCase__ = dataset.shuffle(seed=args.seed ) UpperCamelCase__ = os.path.join(a__ , f"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(a__ , index=a__ ) else: dataset.to_json(a__ ) def _UpperCamelCase (a__ :Union[str, Any] , a__ :Any , a__ :Optional[int] , a__ :Union[str, Any] , **a__ :Union[str, Any] ): """simple docstring""" UpperCamelCase__ = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() UpperCamelCase__ = STModelArguments(model_name_or_path=a__ ) UpperCamelCase__ = STDataArguments(train_file=a__ , infer_file=a__ ) UpperCamelCase__ = STTrainingArguments(output_dir=a__ ) UpperCamelCase__ = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(a__ ).items(): setattr(a__ , a__ , a__ ) for key, value in kwargs.items(): if hasattr(a__ , a__ ): setattr(a__ , a__ , a__ ) # Sanity checks UpperCamelCase__ = {} UpperCamelCase__ = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None UpperCamelCase__ = args.train_file UpperCamelCase__ = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None UpperCamelCase__ = args.eval_file for key in data_files: UpperCamelCase__ = data_files[key].split(""".""" )[-1] assert extension in ["csv", "json"], f"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: UpperCamelCase__ = extension else: assert extension == args.data_file_extension, f"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), f"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("""Creating the initial data directory for self-training...""" ) UpperCamelCase__ = f"""{args.output_dir}/self-train_iter-{{}}""".format UpperCamelCase__ = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=a__ ) os.makedirs(a__ , exist_ok=a__ ) accelerator.wait_for_everyone() UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = 0 UpperCamelCase__ = False # Show the progress bar UpperCamelCase__ = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): UpperCamelCase__ = data_dir_format(a__ ) assert os.path.exists(a__ ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 UpperCamelCase__ = os.path.join(a__ , """stage-1""" ) UpperCamelCase__ = { """accelerator""": accelerator, """model_name_or_path""": args.model_name_or_path, """cache_dir""": args.cache_dir, """do_train""": True, """train_file""": data_files["""train"""] if iteration == 0 else data_files["""train_pseudo"""], """do_eval""": True if args.eval_file is not None else False, """eval_file""": data_files["""eval"""], """do_predict""": True, """infer_file""": data_files["""infer"""], """task_name""": args.task_name, """label_list""": args.label_list, """output_dir""": current_output_dir, """eval_metric""": args.eval_metric, """evaluation_strategy""": args.evaluation_strategy, """early_stopping_patience""": args.early_stopping_patience, """early_stopping_threshold""": args.early_stopping_threshold, """seed""": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(a__ , a__ ): arguments_dict.update({key: value} ) UpperCamelCase__ = os.path.join(a__ , """best-checkpoint""" , a__ ) if os.path.exists(a__ ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.""" , a__ , a__ , ) else: logger.info("""***** Running self-training: iteration: %d, stage: 1 *****""" , a__ ) finetune(**a__ ) accelerator.wait_for_everyone() assert os.path.exists(a__ ) logger.info("""Self-training job completed: iteration: %d, stage: 1.""" , a__ ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data UpperCamelCase__ = os.path.join(a__ , """best-checkpoint""" ) UpperCamelCase__ = os.path.join(a__ , """stage-2""" ) # Update arguments_dict UpperCamelCase__ = model_path UpperCamelCase__ = data_files["""train"""] UpperCamelCase__ = current_output_dir UpperCamelCase__ = os.path.join(a__ , """best-checkpoint""" , a__ ) if os.path.exists(a__ ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.""" , a__ , a__ , ) else: logger.info("""***** Running self-training: iteration: %d, stage: 2 *****""" , a__ ) finetune(**a__ ) accelerator.wait_for_everyone() assert os.path.exists(a__ ) logger.info("""Self-training job completed: iteration: %d, stage: 2.""" , a__ ) UpperCamelCase__ = iteration UpperCamelCase__ = data_dir_format(iteration + 1 ) UpperCamelCase__ = AutoConfig.from_pretrained(os.path.join(a__ , """best-checkpoint""" ) ) UpperCamelCase__ = config.idalabel UpperCamelCase__ = os.path.join(a__ , """eval_results_best-checkpoint.json""" ) UpperCamelCase__ = os.path.join(a__ , """test_results_best-checkpoint.json""" ) assert os.path.exists(a__ ) with open(a__ , """r""" ) as f: UpperCamelCase__ = float(json.load(a__ )[args.eval_metric] ) UpperCamelCase__ = os.path.join(a__ , """infer_output_best-checkpoint.csv""" ) assert os.path.exists(a__ ) # Loading the dataset from local csv or json files. UpperCamelCase__ = load_dataset(args.data_file_extension , data_files={"""data""": data_files["""infer"""]} )["""data"""] UpperCamelCase__ = load_dataset("""csv""" , data_files={"""data""": infer_output_file} )["""data"""] if accelerator.is_main_process: os.makedirs(a__ , exist_ok=a__ ) shutil.copy(a__ , os.path.join(a__ , f"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(a__ ): shutil.copy(a__ , os.path.join(a__ , f"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(a__ , a__ , a__ , a__ , a__ , a__ ) accelerator.wait_for_everyone() UpperCamelCase__ = os.path.join(a__ , f"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: UpperCamelCase__ = eval_result if best_iteration is None: UpperCamelCase__ = new_iteration UpperCamelCase__ = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: UpperCamelCase__ = new_iteration UpperCamelCase__ = new_eval_result UpperCamelCase__ = 0 else: if new_eval_result == best_eval_result: UpperCamelCase__ = new_iteration UpperCamelCase__ = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: UpperCamelCase__ = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("""Best iteration: %d""" , a__ ) logger.info("""Best evaluation result: %s = %f""" , args.eval_metric , a__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(a__ , f"""eval_results_iter-{iteration}.json""" ) , os.path.join(a__ , """eval_results_best-iteration.json""" ) , ) else: # Assume that the last iteration is the best logger.info("""Best iteration: %d""" , args.max_selftrain_iterations - 1 ) logger.info("""Best evaluation result: %s = %f""" , args.eval_metric , a__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(a__ , f"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) , os.path.join(a__ , """eval_results_best-iteration.json""" ) , )
87
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Union[str, Any] = {"""vocab_file""": """sentencepiece.bpe.model"""} SCREAMING_SNAKE_CASE : int = { """vocab_file""": { """camembert-base""": """https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model""", } } SCREAMING_SNAKE_CASE : Union[str, Any] = { """camembert-base""": 512, } SCREAMING_SNAKE_CASE : List[Any] = """▁""" class _UpperCAmelCase ( __snake_case ): '''simple docstring''' lowerCamelCase__ =VOCAB_FILES_NAMES lowerCamelCase__ =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ =['input_ids', 'attention_mask'] def __init__(self , a_ , a_="<s>" , a_="</s>" , a_="</s>" , a_="<s>" , a_="<unk>" , a_="<pad>" , a_="<mask>" , a_=["<s>NOTUSED", "</s>NOTUSED"] , a_ = None , **a_ , ): '''simple docstring''' __snake_case : List[Any] = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else mask_token __snake_case : str = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , cls_token=a_ , pad_token=a_ , mask_token=a_ , additional_special_tokens=a_ , sp_model_kwargs=self.sp_model_kwargs , **a_ , ) __snake_case : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(a_ ) ) __snake_case : Optional[Any] = vocab_file # HACK: These tokens were added by fairseq but don't seem to be actually used when duplicated in the actual # sentencepiece vocabulary (this is the case for <s> and </s> __snake_case : List[Any] = {'''<s>NOTUSED''': 0, '''<pad>''': 1, '''</s>NOTUSED''': 2, '''<unk>''': 3} __snake_case : str = len(self.fairseq_tokens_to_ids ) __snake_case : List[Any] = len(self.sp_model ) + len(self.fairseq_tokens_to_ids ) __snake_case : List[str] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def SCREAMING_SNAKE_CASE (self , a_ , a_ = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __snake_case : Optional[Any] = [self.cls_token_id] __snake_case : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE (self , a_ , a_ = None , a_ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a_ , token_ids_a=a_ , already_has_special_tokens=a_ ) if token_ids_a is None: return [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1, 1] + ([0] * len(a_ )) + [1] def SCREAMING_SNAKE_CASE (self , a_ , a_ = None ): '''simple docstring''' __snake_case : Optional[int] = [self.sep_token_id] __snake_case : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return len(self.fairseq_tokens_to_ids ) + len(self.sp_model ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Dict = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' return self.sp_model.encode(a_ , out_type=a_ ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] elif self.sp_model.PieceToId(a_ ) == 0: # Convert sentence piece unk token to fairseq unk token index return self.unk_token_id return self.fairseq_offset + self.sp_model.PieceToId(a_ ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def SCREAMING_SNAKE_CASE (self , a_ ): '''simple docstring''' __snake_case : Tuple = [] __snake_case : Dict = '''''' __snake_case : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(a_ ) + token __snake_case : Tuple = True __snake_case : Any = [] else: current_sub_tokens.append(a_ ) __snake_case : int = False out_string += self.sp_model.decode(a_ ) return out_string.strip() def __getstate__(self ): '''simple docstring''' __snake_case : Dict = self.__dict__.copy() __snake_case : Union[str, Any] = None return state def __setstate__(self , a_ ): '''simple docstring''' __snake_case : int = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __snake_case : Optional[Any] = {} __snake_case : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE (self , a_ , a_ = None ): '''simple docstring''' if not os.path.isdir(a_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __snake_case : List[Any] = os.path.join( a_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a_ ) elif not os.path.isfile(self.vocab_file ): with open(a_ , '''wb''' ) as fi: __snake_case : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(a_ ) return (out_vocab_file,)
102
def _SCREAMING_SNAKE_CASE ( a ) -> Tuple: __A , __A : Optional[Any] = [], [] while len(a ) > 1: __A , __A : Any = min(a ), max(a ) start.append(a ) end.append(a ) collection.remove(a ) collection.remove(a ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCAmelCase : int = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase : Dict = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
280
0
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def snake_case_(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" _snake_case = AlbertConfig.from_json_file(_UpperCamelCase ) print(F"""Building PyTorch model from configuration: {config}""" ) _snake_case = AlbertForPreTraining(_UpperCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_albert(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _UpperCamelCase ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __A = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
367
from math import factorial def snake_case_(_UpperCamelCase , _UpperCamelCase ) -> int: """simple docstring""" if n < k or k < 0: raise ValueError('''Please enter positive integers for n and k where n >= k''' ) return factorial(_UpperCamelCase ) // (factorial(_UpperCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( '''If a class of 40 students must be arranged into groups of''', f'''4 for group projects, there are {combinations(40, 4)} ways''', '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f'''are {combinations(10, 3)} ways that first, second and''', '''third place can be awarded.''', )
278
0
"""simple docstring""" import datasets lowerCAmelCase_ = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' lowerCAmelCase_ = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' lowerCAmelCase_ = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> Tuple: return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): '''simple docstring''' def UpperCAmelCase ( self : Optional[int] ) -> List[Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32''' ), '''references''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32''' ), } ) ,codebase_urls=[] ,reference_urls=[] ,format='''numpy''' ,) def UpperCAmelCase ( self : List[Any] ,_snake_case : Union[str, Any] ,_snake_case : Optional[int] ) -> Tuple: """simple docstring""" return {"accuracy": simple_accuracy(_snake_case ,_snake_case )}
16
import cva import numpy as np class a : """simple docstring""" def __init__( self : Optional[Any] , __lowercase : float , __lowercase : int ) -> List[Any]: if k in (0.04, 0.06): __UpperCAmelCase : int = k __UpperCAmelCase : List[str] = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : Union[str, Any] ) -> str: return str(self.k ) def UpperCAmelCase ( self : Optional[int] , __lowercase : str ) -> tuple[cva.Mat, list[list[int]]]: __UpperCAmelCase : Optional[Any] = cva.imread(__lowercase , 0 ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = img.shape __UpperCAmelCase : list[list[int]] = [] __UpperCAmelCase : Any = img.copy() __UpperCAmelCase : Tuple = cva.cvtColor(__lowercase , cva.COLOR_GRAY2RGB ) __UpperCAmelCase , __UpperCAmelCase : List[str] = np.gradient(__lowercase ) __UpperCAmelCase : int = dx**2 __UpperCAmelCase : Any = dy**2 __UpperCAmelCase : Any = dx * dy __UpperCAmelCase : Optional[int] = 0.04 __UpperCAmelCase : List[str] = self.window_size // 2 for y in range(__lowercase , h - offset ): for x in range(__lowercase , w - offset ): __UpperCAmelCase : Tuple = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __UpperCAmelCase : List[Any] = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __UpperCAmelCase : Optional[int] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __UpperCAmelCase : int = (wxx * wyy) - (wxy**2) __UpperCAmelCase : Any = wxx + wyy __UpperCAmelCase : Optional[Any] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": a : int = HarrisCorner(0.04, 3) a ,a : Any = edge_detect.detect("path_to_image") cva.imwrite("detect.png", color_img)
114
0
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __A = logging.get_logger(__name__) __A = { 'microsoft/deberta-v2-xlarge': 'https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json', 'microsoft/deberta-v2-xxlarge': 'https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json', 'microsoft/deberta-v2-xlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json' ), 'microsoft/deberta-v2-xxlarge-mnli': ( 'https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json' ), } class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' lowerCamelCase = '''deberta-v2''' def __init__( self , __UpperCAmelCase=12_81_00 , __UpperCAmelCase=15_36 , __UpperCAmelCase=24 , __UpperCAmelCase=24 , __UpperCAmelCase=61_44 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=5_12 , __UpperCAmelCase=0 , __UpperCAmelCase=0.0_2 , __UpperCAmelCase=1e-7 , __UpperCAmelCase=False , __UpperCAmelCase=-1 , __UpperCAmelCase=0 , __UpperCAmelCase=True , __UpperCAmelCase=None , __UpperCAmelCase=0 , __UpperCAmelCase="gelu" , **__UpperCAmelCase , ) -> Tuple: super().__init__(**__UpperCAmelCase ) _lowerCAmelCase =hidden_size _lowerCAmelCase =num_hidden_layers _lowerCAmelCase =num_attention_heads _lowerCAmelCase =intermediate_size _lowerCAmelCase =hidden_act _lowerCAmelCase =hidden_dropout_prob _lowerCAmelCase =attention_probs_dropout_prob _lowerCAmelCase =max_position_embeddings _lowerCAmelCase =type_vocab_size _lowerCAmelCase =initializer_range _lowerCAmelCase =relative_attention _lowerCAmelCase =max_relative_positions _lowerCAmelCase =pad_token_id _lowerCAmelCase =position_biased_input # Backwards compatibility if type(__UpperCAmelCase ) == str: _lowerCAmelCase =[x.strip() for x in pos_att_type.lower().split("""|""" )] _lowerCAmelCase =pos_att_type _lowerCAmelCase =vocab_size _lowerCAmelCase =layer_norm_eps _lowerCAmelCase =kwargs.get("""pooler_hidden_size""" , __UpperCAmelCase ) _lowerCAmelCase =pooler_dropout _lowerCAmelCase =pooler_hidden_act class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' @property def _lowerCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _lowerCAmelCase ={0: """batch""", 1: """choice""", 2: """sequence"""} else: _lowerCAmelCase ={0: """batch""", 1: """sequence"""} if self._config.type_vocab_size > 0: return OrderedDict( [("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis)] ) else: return OrderedDict([("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis)] ) @property def _lowerCAmelCase ( self ) -> int: return 12 def _lowerCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = 3 , __UpperCAmelCase = 40 , __UpperCAmelCase = 40 , __UpperCAmelCase = None , ) -> Mapping[str, Any]: _lowerCAmelCase =super().generate_dummy_inputs(preprocessor=__UpperCAmelCase , framework=__UpperCAmelCase ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
341
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
1
"""simple docstring""" from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _SCREAMING_SNAKE_CASE( UpperCamelCase_ ): SCREAMING_SNAKE_CASE_ : str = ['''image_processor''', '''tokenizer'''] SCREAMING_SNAKE_CASE_ : str = '''Pix2StructImageProcessor''' SCREAMING_SNAKE_CASE_ : List[Any] = ('''T5Tokenizer''', '''T5TokenizerFast''') def __init__( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[str] = False super().__init__(UpperCamelCase__ ,UpperCamelCase__ ) def __call__( self ,SCREAMING_SNAKE_CASE__=None ,SCREAMING_SNAKE_CASE__ = None ,SCREAMING_SNAKE_CASE__ = True ,SCREAMING_SNAKE_CASE__ = False ,SCREAMING_SNAKE_CASE__ = None ,SCREAMING_SNAKE_CASE__ = None ,SCREAMING_SNAKE_CASE__ = 20_48 ,SCREAMING_SNAKE_CASE__ = 0 ,SCREAMING_SNAKE_CASE__ = None ,SCREAMING_SNAKE_CASE__ = None ,SCREAMING_SNAKE_CASE__ = False ,SCREAMING_SNAKE_CASE__ = False ,SCREAMING_SNAKE_CASE__ = False ,SCREAMING_SNAKE_CASE__ = False ,SCREAMING_SNAKE_CASE__ = False ,SCREAMING_SNAKE_CASE__ = True ,SCREAMING_SNAKE_CASE__ = None ,**SCREAMING_SNAKE_CASE__ ,) -> Any: """simple docstring""" if images is None and text is None: raise ValueError('''You have to specify either images or text.''' ) # Get only text if images is None and not self.image_processor.is_vqa: __SCREAMING_SNAKE_CASE :int = self.tokenizer __SCREAMING_SNAKE_CASE :Union[str, Any] = self.tokenizer( text=UpperCamelCase__ ,add_special_tokens=UpperCamelCase__ ,padding=UpperCamelCase__ ,truncation=UpperCamelCase__ ,max_length=UpperCamelCase__ ,stride=UpperCamelCase__ ,pad_to_multiple_of=UpperCamelCase__ ,return_attention_mask=UpperCamelCase__ ,return_overflowing_tokens=UpperCamelCase__ ,return_special_tokens_mask=UpperCamelCase__ ,return_offsets_mapping=UpperCamelCase__ ,return_token_type_ids=UpperCamelCase__ ,return_length=UpperCamelCase__ ,verbose=UpperCamelCase__ ,return_tensors=UpperCamelCase__ ,**UpperCamelCase__ ,) return text_encoding if not self.image_processor.is_vqa: # add pixel_values __SCREAMING_SNAKE_CASE :int = self.image_processor( UpperCamelCase__ ,return_tensors=UpperCamelCase__ ,max_patches=UpperCamelCase__ ,**UpperCamelCase__ ) else: # add pixel_values and bbox __SCREAMING_SNAKE_CASE :Optional[Any] = self.image_processor( UpperCamelCase__ ,return_tensors=UpperCamelCase__ ,max_patches=UpperCamelCase__ ,header_text=UpperCamelCase__ ,**UpperCamelCase__ ) if text is not None and not self.image_processor.is_vqa: __SCREAMING_SNAKE_CASE :Optional[int] = self.tokenizer( text=UpperCamelCase__ ,add_special_tokens=UpperCamelCase__ ,padding=UpperCamelCase__ ,truncation=UpperCamelCase__ ,max_length=UpperCamelCase__ ,stride=UpperCamelCase__ ,pad_to_multiple_of=UpperCamelCase__ ,return_attention_mask=UpperCamelCase__ ,return_overflowing_tokens=UpperCamelCase__ ,return_special_tokens_mask=UpperCamelCase__ ,return_offsets_mapping=UpperCamelCase__ ,return_token_type_ids=UpperCamelCase__ ,return_length=UpperCamelCase__ ,verbose=UpperCamelCase__ ,return_tensors=UpperCamelCase__ ,**UpperCamelCase__ ,) if "attention_mask" in text_encoding: __SCREAMING_SNAKE_CASE :str = text_encoding.pop('''attention_mask''' ) if "input_ids" in text_encoding: __SCREAMING_SNAKE_CASE :List[str] = text_encoding.pop('''input_ids''' ) else: __SCREAMING_SNAKE_CASE :List[Any] = None if text_encoding is not None: encoding_image_processor.update(UpperCamelCase__ ) return encoding_image_processor def _UpperCamelCase ( self ,*SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) -> List[Any]: """simple docstring""" return self.tokenizer.batch_decode(*UpperCamelCase__ ,**UpperCamelCase__ ) def _UpperCamelCase ( self ,*SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ) -> Any: """simple docstring""" return self.tokenizer.decode(*UpperCamelCase__ ,**UpperCamelCase__ ) @property def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = self.tokenizer.model_input_names __SCREAMING_SNAKE_CASE :Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
191
import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params __UpperCamelCase : Any = getLogger(__name__) __UpperCamelCase : int = 'cuda' if torch.cuda.is_available() else 'cpu' def A ( _lowercase , _lowercase , _lowercase , _lowercase = 8 , _lowercase = DEFAULT_DEVICE , _lowercase=False , _lowercase="summarization" , _lowercase=None , **_lowercase , ): SCREAMING_SNAKE_CASE : List[str] = Path(_lowercase ).open('''w''' , encoding='''utf-8''' ) SCREAMING_SNAKE_CASE : int = str(_lowercase ) SCREAMING_SNAKE_CASE : Any = AutoModelForSeqaSeqLM.from_pretrained(_lowercase ).to(_lowercase ) if fpaa: SCREAMING_SNAKE_CASE : Dict = model.half() SCREAMING_SNAKE_CASE : Optional[Any] = AutoTokenizer.from_pretrained(_lowercase ) logger.info(f"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. SCREAMING_SNAKE_CASE : str = time.time() # update config with task specific params use_task_specific_params(_lowercase , _lowercase ) if prefix is None: SCREAMING_SNAKE_CASE : Optional[int] = prefix or getattr(model.config , '''prefix''' , '''''' ) or '''''' for examples_chunk in tqdm(list(chunks(_lowercase , _lowercase ) ) ): SCREAMING_SNAKE_CASE : Union[str, Any] = [prefix + text for text in examples_chunk] SCREAMING_SNAKE_CASE : Dict = tokenizer(_lowercase , return_tensors='''pt''' , truncation=_lowercase , padding='''longest''' ).to(_lowercase ) SCREAMING_SNAKE_CASE : str = model.generate( input_ids=batch.input_ids , attention_mask=batch.attention_mask , **_lowercase , ) SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.batch_decode(_lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() SCREAMING_SNAKE_CASE : Tuple = int(time.time() - start_time ) # seconds SCREAMING_SNAKE_CASE : str = len(_lowercase ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs , 4 )} def A ( ): return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def A ( _lowercase=True ): SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() parser.add_argument('''model_name''' , type=_lowercase , help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' , type=_lowercase , help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' , type=_lowercase , help='''where to save summaries''' ) parser.add_argument('''--reference_path''' , type=_lowercase , required=_lowercase , help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' , type=_lowercase , required=_lowercase , default='''metrics.json''' , help='''where to save metrics''' ) parser.add_argument('''--device''' , type=_lowercase , required=_lowercase , default=_lowercase , help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' , type=_lowercase , required=_lowercase , default=_lowercase , help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' , type=_lowercase , default='''summarization''' , help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' , type=_lowercase , default=8 , required=_lowercase , help='''batch size''' ) parser.add_argument( '''--n_obs''' , type=_lowercase , default=-1 , required=_lowercase , help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' , action='''store_true''' ) parser.add_argument('''--dump-args''' , action='''store_true''' , help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' , nargs='''?''' , type=_lowercase , const=datetime_now() , help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) , ) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_known_args() SCREAMING_SNAKE_CASE : Optional[Any] = parse_numeric_n_bool_cl_kwargs(_lowercase ) if parsed_args and verbose: print(f"""parsed the following generate kwargs: {parsed_args}""" ) SCREAMING_SNAKE_CASE : Union[str, Any] = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: SCREAMING_SNAKE_CASE : Any = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=_lowercase ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(f"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) SCREAMING_SNAKE_CASE : List[str] = generate_summaries_or_translations( _lowercase , args.save_path , args.model_name , batch_size=args.bs , device=args.device , fpaa=args.fpaa , task=args.task , prefix=args.prefix , **_lowercase , ) if args.reference_path is None: return {} # Compute scores SCREAMING_SNAKE_CASE : Dict = calculate_bleu if '''translation''' in args.task else calculate_rouge SCREAMING_SNAKE_CASE : Union[str, Any] = [x.rstrip() for x in open(args.save_path ).readlines()] SCREAMING_SNAKE_CASE : Optional[int] = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(_lowercase )] SCREAMING_SNAKE_CASE : dict = score_fn(_lowercase , _lowercase ) scores.update(_lowercase ) if args.dump_args: scores.update(_lowercase ) if args.info: SCREAMING_SNAKE_CASE : Tuple = args.info if verbose: print(_lowercase ) if args.score_path is not None: json.dump(_lowercase , open(args.score_path , '''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
182
0
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging __UpperCAmelCase = logging.get_logger(__name__) def __lowerCamelCase ( __magic_name__ : str , __magic_name__ : int ): a__: Union[str, Any] =set() a__: Union[str, Any] =[] def parse_line(__magic_name__ : Tuple ): for line in fp: if isinstance(__magic_name__ , __magic_name__ ): a__: str =line.decode("UTF-8" ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(" " ): # process a single warning and move it to `selected_warnings`. if len(__magic_name__ ) > 0: a__: Optional[int] ="\n".join(__magic_name__ ) # Only keep the warnings specified in `targets` if any(F": {x}: " in warning for x in targets ): selected_warnings.add(__magic_name__ ) buffer.clear() continue else: a__: List[str] =line.strip() buffer.append(__magic_name__ ) if from_gh: for filename in os.listdir(__magic_name__ ): a__: Dict =os.path.join(__magic_name__ , __magic_name__ ) if not os.path.isdir(__magic_name__ ): # read the file if filename != "warnings.txt": continue with open(__magic_name__ ) as fp: parse_line(__magic_name__ ) else: try: with zipfile.ZipFile(__magic_name__ ) as z: for filename in z.namelist(): if not os.path.isdir(__magic_name__ ): # read the file if filename != "warnings.txt": continue with z.open(__magic_name__ ) as fp: parse_line(__magic_name__ ) except Exception: logger.warning( F"{artifact_path} is either an invalid zip file or something else wrong. This file is skipped." ) return selected_warnings def __lowerCamelCase ( __magic_name__ : Dict , __magic_name__ : Union[str, Any] ): a__: Tuple =set() a__: Union[str, Any] =[os.path.join(__magic_name__ , __magic_name__ ) for p in os.listdir(__magic_name__ ) if (p.endswith(".zip" ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(__magic_name__ , __magic_name__ ) ) return selected_warnings if __name__ == "__main__": def __lowerCamelCase ( __magic_name__ : int ): return values.split("," ) __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('''--workflow_run_id''', type=str, required=True, help='''A GitHub Actions workflow run id.''') parser.add_argument( '''--output_dir''', type=str, required=True, help='''Where to store the downloaded artifacts and other result files.''', ) parser.add_argument('''--token''', default=None, type=str, help='''A token that has actions:read permission.''') # optional parameters parser.add_argument( '''--targets''', default='''DeprecationWarning,UserWarning,FutureWarning''', type=list_str, help='''Comma-separated list of target warning(s) which we want to extract.''', ) parser.add_argument( '''--from_gh''', action='''store_true''', help='''If running from a GitHub action workflow and collecting warnings from its artifacts.''', ) __UpperCAmelCase = parser.parse_args() __UpperCAmelCase = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links __UpperCAmelCase = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, '''artifacts.json'''), '''w''', encoding='''UTF-8''') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print('''=''' * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts __UpperCAmelCase = extract_warnings(args.output_dir, args.targets) __UpperCAmelCase = sorted(selected_warnings) with open(os.path.join(args.output_dir, '''selected_warnings.json'''), '''w''', encoding='''UTF-8''') as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
358
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __UpperCAmelCase = 2_00 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __UpperCAmelCase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __UpperCAmelCase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 10_00)) def __lowerCamelCase ( __magic_name__ : str , __magic_name__ : str ): a__: int =len([g for position, g in enumerate(__magic_name__ ) if g == main_target[position]] ) return (item, float(__magic_name__ )) def __lowerCamelCase ( __magic_name__ : str , __magic_name__ : str ): a__: Any =random.randint(0 , len(__magic_name__ ) - 1 ) a__: Tuple =parent_a[:random_slice] + parent_a[random_slice:] a__: List[str] =parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def __lowerCamelCase ( __magic_name__ : str , __magic_name__ : list[str] ): a__: str =list(__magic_name__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: a__: Union[str, Any] =random.choice(__magic_name__ ) return "".join(__magic_name__ ) def __lowerCamelCase ( __magic_name__ : tuple[str, float] , __magic_name__ : list[tuple[str, float]] , __magic_name__ : list[str] , ): a__: List[Any] =[] # Generate more children proportionally to the fitness score. a__: Dict =int(parent_a[1] * 100 ) + 1 a__: Tuple =10 if child_n >= 10 else child_n for _ in range(__magic_name__ ): a__: List[str] =population_score[random.randint(0 , __magic_name__ )][0] a__ , a__: Dict =crossover(parent_a[0] , __magic_name__ ) # Append new string to the population list. pop.append(mutate(__magic_name__ , __magic_name__ ) ) pop.append(mutate(__magic_name__ , __magic_name__ ) ) return pop def __lowerCamelCase ( __magic_name__ : str , __magic_name__ : list[str] , __magic_name__ : bool = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: a__: Any =F"{N_POPULATION} must be bigger than {N_SELECTED}" raise ValueError(__magic_name__ ) # Verify that the target contains no genes besides the ones inside genes variable. a__: int =sorted({c for c in target if c not in genes} ) if not_in_genes_list: a__: str =F"{not_in_genes_list} is not in genes list, evolution cannot converge" raise ValueError(__magic_name__ ) # Generate random starting population. a__: Tuple =[] for _ in range(__magic_name__ ): population.append("".join([random.choice(__magic_name__ ) for i in range(len(__magic_name__ ) )] ) ) # Just some logs to know what the algorithms is doing. a__ , a__: Any =0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__magic_name__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. a__: Dict =[evaluate(__magic_name__ , __magic_name__ ) for item in population] # Check if there is a matching evolution. a__: Any =sorted(__magic_name__ , key=lambda __magic_name__ : x[1] , reverse=__magic_name__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F"\nGeneration: {generation}" F"\nTotal Population:{total_population}" F"\nBest score: {population_score[0][1]}" F"\nBest string: {population_score[0][0]}" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. a__: Optional[int] =population[: int(N_POPULATION / 3 )] population.clear() population.extend(__magic_name__ ) # Normalize population score to be between 0 and 1. a__: List[str] =[ (item, score / len(__magic_name__ )) for item, score in population_score ] # This is selection for i in range(__magic_name__ ): population.extend(select(population_score[int(__magic_name__ )] , __magic_name__ , __magic_name__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__magic_name__ ) > N_POPULATION: break if __name__ == "__main__": __UpperCAmelCase = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __UpperCAmelCase = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = basic(target_str, genes_list) print( f"""\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}""" )
42
0
"""simple docstring""" from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class _A : """simple docstring""" def __init__( self : str , __UpperCAmelCase : Optional[Any] , ): a : List[str] = parent a : Optional[Any] = 13 a : Dict = 7 a : Dict = 30 a : Optional[int] = self.seq_length + self.mem_len a : Optional[int] = 15 a : Dict = True a : List[str] = True a : Union[str, Any] = 99 a : Any = [10, 50, 80] a : Dict = 32 a : Dict = 32 a : Optional[int] = 4 a : List[Any] = 8 a : List[Any] = 128 a : Optional[int] = 2 a : List[Any] = 2 a : Tuple = None a : str = 1 a : Dict = 0 a : Any = 3 a : Any = self.vocab_size - 1 a : Optional[int] = 0.01 def __snake_case ( self : Union[str, Any]): a : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a : Any = None if self.use_labels: a : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a : Tuple = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def __snake_case ( self : List[str]): random.seed(self.seed) tf.random.set_seed(self.seed) def __snake_case ( self : List[str] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Any , __UpperCAmelCase : int): a : List[str] = TFTransfoXLModel(__UpperCAmelCase) a , a : List[str] = model(__UpperCAmelCase).to_tuple() a : Optional[Any] = {"input_ids": input_ids_a, "mems": mems_a} a , a : Union[str, Any] = model(__UpperCAmelCase).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __snake_case ( self : int , __UpperCAmelCase : str , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, Any]): a : List[Any] = TFTransfoXLLMHeadModel(__UpperCAmelCase) a , a : Optional[Any] = model(__UpperCAmelCase).to_tuple() a : Optional[int] = {"input_ids": input_ids_a, "labels": lm_labels} a , a : Optional[int] = model(__UpperCAmelCase).to_tuple() a , a : Optional[int] = model([input_ids_a, mems_a]).to_tuple() a : Tuple = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} a , a : List[str] = model(__UpperCAmelCase).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __snake_case ( self : List[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : List[str]): a : Dict = TFTransfoXLForSequenceClassification(__UpperCAmelCase) a : int = model(__UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def __snake_case ( self : Union[str, Any]): a : Optional[Any] = self.prepare_config_and_inputs() ((a) , (a) , (a) , (a)) : List[Any] = config_and_inputs a : List[str] = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class _A ( _a ,_a ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : Any = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) UpperCAmelCase : List[Any] = () if is_tf_available() else () UpperCAmelCase : Tuple = ( { """feature-extraction""": TFTransfoXLModel, """text-classification""": TFTransfoXLForSequenceClassification, """text-generation""": TFTransfoXLLMHeadModel, """zero-shot""": TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented UpperCAmelCase : Any = False UpperCAmelCase : int = False UpperCAmelCase : Dict = False UpperCAmelCase : Any = False def __snake_case ( self : List[str] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[int]): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def __snake_case ( self : Any): a : Any = TFTransfoXLModelTester(self) a : Any = ConfigTester(self , config_class=__UpperCAmelCase , d_embed=37) def __snake_case ( self : int): self.config_tester.run_common_tests() def __snake_case ( self : int): self.model_tester.set_seed() a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*__UpperCAmelCase) def __snake_case ( self : str): self.model_tester.set_seed() a : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*__UpperCAmelCase) def __snake_case ( self : Optional[int]): a : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*__UpperCAmelCase) def __snake_case ( self : Tuple): a , a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() a : Optional[Any] = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: a : Optional[int] = model_class(__UpperCAmelCase) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer) if model_class in list_other_models_with_output_ebd: a : Optional[int] = model.get_output_embeddings() assert isinstance(__UpperCAmelCase , tf.keras.layers.Layer) a : List[str] = model.get_bias() assert name is None else: a : List[str] = model.get_output_embeddings() assert x is None a : Optional[int] = model.get_bias() assert name is None def __snake_case ( self : Optional[Any]): # TODO JP: Make TransfoXL XLA compliant pass @slow def __snake_case ( self : str): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a : Tuple = TFTransfoXLModel.from_pretrained(__UpperCAmelCase) self.assertIsNotNone(__UpperCAmelCase) @unittest.skip(reason="This model doesn't play well with fit() due to not returning a single loss.") def __snake_case ( self : Tuple): pass @require_tf class _A ( unittest.TestCase ): """simple docstring""" @unittest.skip("Skip test until #12651 is resolved.") @slow def __snake_case ( self : List[str]): a : int = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103") # fmt: off a : Tuple = tf.convert_to_tensor([[33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0]] , dtype=tf.intaa) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off a : Union[str, Any] = [33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0,33,1,1857,2,1,1009,4,1109,11739,4762,358,5,25,245,28,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> a : Optional[int] = model.generate(__UpperCAmelCase , max_length=200 , do_sample=__UpperCAmelCase) self.assertListEqual(output_ids[0].numpy().tolist() , __UpperCAmelCase)
40
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = tempfile.mkdtemp() # fmt: off lowerCamelCase__ : int = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest"""] # fmt: on lowerCamelCase__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) lowerCamelCase__ : Tuple = { """do_resize""": True, """size""": {"""height""": 18, """width""": 18}, """do_normalize""": True, """image_mean""": [0.5, 0.5, 0.5], """image_std""": [0.5, 0.5, 0.5], } lowerCamelCase__ : Tuple = os.path.join(self.tmpdirname , UpperCamelCase__ ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: str , **UpperCamelCase__: List[str] ): return BertTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: int , **UpperCamelCase__: Tuple ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] ): shutil.rmtree(self.tmpdirname ) def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowerCamelCase__ : Tuple = [Image.fromarray(np.moveaxis(UpperCamelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self: int ): lowerCamelCase__ : Optional[Any] = self.get_tokenizer() lowerCamelCase__ : Dict = self.get_image_processor() lowerCamelCase__ : Optional[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : Dict = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowerCamelCase__ : int = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) lowerCamelCase__ : List[Any] = self.get_image_processor(do_normalize=UpperCamelCase__ , padding_value=1.0 ) lowerCamelCase__ : Tuple = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCamelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : Any = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = self.prepare_image_inputs() lowerCamelCase__ : List[str] = image_processor(UpperCamelCase__ , return_tensors="""np""" ) lowerCamelCase__ : Optional[Any] = processor(images=UpperCamelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : List[Any] = """lower newer""" lowerCamelCase__ : Union[str, Any] = processor(text=UpperCamelCase__ ) lowerCamelCase__ : Any = tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Optional[Any] = self.get_image_processor() lowerCamelCase__ : List[Any] = self.get_tokenizer() lowerCamelCase__ : List[Any] = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Any = """lower newer""" lowerCamelCase__ : Dict = self.prepare_image_inputs() lowerCamelCase__ : Tuple = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """token_type_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with self.assertRaises(UpperCamelCase__ ): processor() def lowerCamelCase_ ( self: int ): lowerCamelCase__ : List[str] = self.get_image_processor() lowerCamelCase__ : List[str] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCamelCase__ : Union[str, Any] = processor.batch_decode(UpperCamelCase__ ) lowerCamelCase__ : Dict = tokenizer.batch_decode(UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Any = self.get_image_processor() lowerCamelCase__ : Union[str, Any] = self.get_tokenizer() lowerCamelCase__ : int = VisionTextDualEncoderProcessor(tokenizer=UpperCamelCase__ , image_processor=UpperCamelCase__ ) lowerCamelCase__ : Optional[Any] = """lower newer""" lowerCamelCase__ : str = self.prepare_image_inputs() lowerCamelCase__ : int = processor(text=UpperCamelCase__ , images=UpperCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
41
0
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union UpperCAmelCase = TypeVar('''T''') UpperCAmelCase = Union[List[T], Tuple[T, ...]] UpperCAmelCase = Union[T, List[T], Dict[str, T]] UpperCAmelCase = Union[str, bytes, os.PathLike]
172
"""simple docstring""" def lowerCamelCase (a_ :list , a_ :list , a_ :int , a_ :int , a_ :int) -> int: if index == number_of_items: return 0 lowercase :Optional[int] = 0 lowercase :str = 0 lowercase :List[str] = knapsack(a_ , a_ , a_ , a_ , index + 1) if weights[index] <= max_weight: lowercase :Any = values[index] + knapsack( a_ , a_ , a_ , max_weight - weights[index] , index + 1) return max(a_ , a_) if __name__ == "__main__": import doctest doctest.testmod()
172
1