code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __a = """.""" # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __a = [ """Assert""", """AssignVariableOp""", """EmptyTensorList""", """MergeV2Checkpoints""", """ReadVariableOp""", """ResourceGather""", """RestoreV2""", """SaveV2""", """ShardedFilename""", """StatefulPartitionedCall""", """StaticRegexFullMatch""", """VarHandleOp""", ] def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: UpperCAmelCase = SavedModel() UpperCAmelCase = [] with open(os.path.join(lowerCAmelCase_ , """utils""" , """tf_ops""" , """onnx.json""" ) ) as f: UpperCAmelCase = json.load(lowerCAmelCase_ )["""opsets"""] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(lowerCAmelCase_ )] ) with open(lowerCAmelCase_ , """rb""" ) as f: saved_model.ParseFromString(f.read() ) UpperCAmelCase = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want UpperCAmelCase = sorted(lowerCAmelCase_ ) UpperCAmelCase = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(lowerCAmelCase_ ) if strict and len(lowerCAmelCase_ ) > 0: raise Exception(F"""Found the following incompatible ops for the opset {opset}:\n""" + incompatible_ops ) elif len(lowerCAmelCase_ ) > 0: print(F"""Found the following incompatible ops for the opset {opset}:""" ) print(*lowerCAmelCase_ , sep="""\n""" ) else: print(F"""The saved model {saved_model_path} can properly be converted with ONNX.""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--saved_model_path""", help="""Path of the saved model to check (the .pb file).""") parser.add_argument( """--opset""", default=12, type=int, help="""The ONNX opset against which the model has to be tested.""" ) parser.add_argument( """--framework""", choices=["""onnx"""], default="""onnx""", help="""Frameworks against which to test the saved model.""" ) parser.add_argument( """--strict""", action="""store_true""", help="""Whether make the checking strict (raise errors) or not (raise warnings)""" ) __a = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """microsoft/beit-base-patch16-224-pt22k""": ( """https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json""" ), # See all BEiT models at https://huggingface.co/models?filter=beit } class __lowercase ( __snake_case ): UpperCamelCase = '''beit''' def __init__( self : Any , __lowerCamelCase : Tuple=8_1_9_2 , __lowerCamelCase : Dict=7_6_8 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=1_2 , __lowerCamelCase : str=3_0_7_2 , __lowerCamelCase : Dict="gelu" , __lowerCamelCase : Optional[int]=0.0 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Tuple=0.02 , __lowerCamelCase : Tuple=1e-1_2 , __lowerCamelCase : Tuple=2_2_4 , __lowerCamelCase : str=1_6 , __lowerCamelCase : Optional[int]=3 , __lowerCamelCase : str=False , __lowerCamelCase : List[str]=False , __lowerCamelCase : int=False , __lowerCamelCase : Union[str, Any]=False , __lowerCamelCase : Dict=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Dict=True , __lowerCamelCase : Optional[Any]=[3, 5, 7, 1_1] , __lowerCamelCase : List[str]=[1, 2, 3, 6] , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : List[Any]=0.4 , __lowerCamelCase : Dict=2_5_6 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : List[str]=False , __lowerCamelCase : List[str]=2_5_5 , **__lowerCamelCase : Optional[int] , ) -> Optional[Any]: """simple docstring""" super().__init__(**__lowerCamelCase ) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = use_mask_token UpperCAmelCase = use_absolute_position_embeddings UpperCAmelCase = use_relative_position_bias UpperCAmelCase = use_shared_relative_position_bias UpperCAmelCase = layer_scale_init_value UpperCAmelCase = drop_path_rate UpperCAmelCase = use_mean_pooling # decode head attributes (semantic segmentation) UpperCAmelCase = out_indices UpperCAmelCase = pool_scales # auxiliary head attributes (semantic segmentation) UpperCAmelCase = use_auxiliary_head UpperCAmelCase = auxiliary_loss_weight UpperCAmelCase = auxiliary_channels UpperCAmelCase = auxiliary_num_convs UpperCAmelCase = auxiliary_concat_input UpperCAmelCase = semantic_loss_ignore_index class __lowercase ( __snake_case ): UpperCamelCase = version.parse('''1.11''' ) @property def _lowercase ( self : Optional[int] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def _lowercase ( self : str ) -> float: """simple docstring""" return 1e-4
627
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
1
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=5 ) ->Dict: # Adapted from https://github.com/pytorch/fairseq/blob/master/fairseq/models/roberta/hub_interface.py assert masked_input.count("""<mask>""" ) == 1 UpperCAmelCase = torch.tensor(tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) ).unsqueeze(0 ) # Batch size 1 UpperCAmelCase = model(lowerCAmelCase_ )[0] # The last hidden-state is the first element of the output tuple UpperCAmelCase = (input_ids.squeeze() == tokenizer.mask_token_id).nonzero().item() UpperCAmelCase = logits[0, masked_index, :] UpperCAmelCase = logits.softmax(dim=0 ) UpperCAmelCase , UpperCAmelCase = prob.topk(k=lowerCAmelCase_ , dim=0 ) UpperCAmelCase = """ """.join( [tokenizer.convert_ids_to_tokens(indices[i].item() ) for i in range(len(lowerCAmelCase_ ) )] ) UpperCAmelCase = tokenizer.mask_token UpperCAmelCase = [] for index, predicted_token_bpe in enumerate(topk_predicted_token_bpe.split(""" """ ) ): UpperCAmelCase = predicted_token_bpe.replace("""\u2581""" , """ """ ) if " {0}".format(lowerCAmelCase_ ) in masked_input: topk_filled_outputs.append( ( masked_input.replace(""" {0}""".format(lowerCAmelCase_ ) , lowerCAmelCase_ ), values[index].item(), predicted_token, ) ) else: topk_filled_outputs.append( ( masked_input.replace(lowerCAmelCase_ , lowerCAmelCase_ ), values[index].item(), predicted_token, ) ) return topk_filled_outputs __a = CamembertTokenizer.from_pretrained("""camembert-base""") __a = CamembertForMaskedLM.from_pretrained("""camembert-base""") model.eval() __a = """Le camembert est <mask> :)""" print(fill_mask(masked_input, model, tokenizer, topk=3))
627
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
1
from ...configuration_utils import PretrainedConfig __a = { """google/tapas-base-finetuned-sqa""": ( """https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json""" ), """google/tapas-base-finetuned-wtq""": ( """https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json""" ), """google/tapas-base-finetuned-wikisql-supervised""": ( """https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json""" ), """google/tapas-base-finetuned-tabfact""": ( """https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json""" ), } class __lowercase ( __snake_case ): UpperCamelCase = '''tapas''' def __init__( self : str , __lowerCamelCase : Optional[Any]=3_0_5_2_2 , __lowerCamelCase : Dict=7_6_8 , __lowerCamelCase : List[Any]=1_2 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : str=3_0_7_2 , __lowerCamelCase : int="gelu" , __lowerCamelCase : Any=0.1 , __lowerCamelCase : int=0.1 , __lowerCamelCase : List[Any]=1_0_2_4 , __lowerCamelCase : Dict=[3, 2_5_6, 2_5_6, 2, 2_5_6, 2_5_6, 1_0] , __lowerCamelCase : Dict=0.02 , __lowerCamelCase : Union[str, Any]=1e-1_2 , __lowerCamelCase : Union[str, Any]=0 , __lowerCamelCase : int=10.0 , __lowerCamelCase : Optional[int]=0 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : Tuple=None , __lowerCamelCase : Any=1.0 , __lowerCamelCase : str=False , __lowerCamelCase : Dict=None , __lowerCamelCase : Any=1.0 , __lowerCamelCase : Tuple=1.0 , __lowerCamelCase : Any=False , __lowerCamelCase : List[str]=False , __lowerCamelCase : List[str]="ratio" , __lowerCamelCase : Tuple=None , __lowerCamelCase : List[Any]=None , __lowerCamelCase : str=6_4 , __lowerCamelCase : List[Any]=3_2 , __lowerCamelCase : str=False , __lowerCamelCase : Dict=True , __lowerCamelCase : int=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : int=True , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : Tuple=None , **__lowerCamelCase : Any , ) -> str: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_sizes UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps # Fine-tuning task hyperparameters UpperCAmelCase = positive_label_weight UpperCAmelCase = num_aggregation_labels UpperCAmelCase = aggregation_loss_weight UpperCAmelCase = use_answer_as_supervision UpperCAmelCase = answer_loss_importance UpperCAmelCase = use_normalized_answer_loss UpperCAmelCase = huber_loss_delta UpperCAmelCase = temperature UpperCAmelCase = aggregation_temperature UpperCAmelCase = use_gumbel_for_cells UpperCAmelCase = use_gumbel_for_aggregation UpperCAmelCase = average_approximation_function UpperCAmelCase = cell_selection_preference UpperCAmelCase = answer_loss_cutoff UpperCAmelCase = max_num_rows UpperCAmelCase = max_num_columns UpperCAmelCase = average_logits_per_cell UpperCAmelCase = select_one_column UpperCAmelCase = allow_empty_column_selection UpperCAmelCase = init_cell_selection_weights_to_zero UpperCAmelCase = reset_position_index_per_cell UpperCAmelCase = disable_per_token_loss # Aggregation hyperparameters UpperCAmelCase = aggregation_labels UpperCAmelCase = no_aggregation_label_index if isinstance(self.aggregation_labels , __lowerCamelCase ): UpperCAmelCase = {int(__lowerCamelCase ): v for k, v in aggregation_labels.items()}
627
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
1
__a = """0.21.0""" from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
1
class __lowercase : def __init__( self : List[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : str , __lowerCamelCase : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = name UpperCAmelCase = value UpperCAmelCase = weight def __repr__( self : Tuple ) -> Optional[Any]: """simple docstring""" return F"""{self.__class__.__name__}({self.name}, {self.value}, {self.weight})""" def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" return self.value def _lowercase ( self : int ) -> Union[str, Any]: """simple docstring""" return self.name def _lowercase ( self : int ) -> Any: """simple docstring""" return self.weight def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" return self.value / self.weight def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Optional[Any]: UpperCAmelCase = sorted(lowerCAmelCase_ , key=lowerCAmelCase_ , reverse=lowerCAmelCase_ ) UpperCAmelCase = [] UpperCAmelCase , UpperCAmelCase = 0.0, 0.0 for i in range(len(lowerCAmelCase_ ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def _UpperCamelCase ( ) ->List[str]: pass if __name__ == "__main__": import doctest doctest.testmod()
627
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
1
import math def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0 ) ->int: UpperCAmelCase = sum(i * i for i in range(1 , n + 1 ) ) UpperCAmelCase = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
627
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
1
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class __lowercase : def __init__( self : List[str] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int]=1_3 , __lowerCamelCase : int=7 , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : List[str]=True , __lowerCamelCase : List[str]=9_9 , __lowerCamelCase : Any=2_4 , __lowerCamelCase : Union[str, Any]=2 , __lowerCamelCase : List[Any]=6 , __lowerCamelCase : str=3_7 , __lowerCamelCase : List[Any]="gelu" , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Optional[Any]=0.1 , __lowerCamelCase : str=5_1_2 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : int=2 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : Any=3 , __lowerCamelCase : int=None , __lowerCamelCase : Optional[int]=1_0_0_0 , ) -> List[Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_input_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = scope UpperCAmelCase = range_bbox def _lowercase ( self : int ) -> List[Any]: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: UpperCAmelCase = bbox[i, j, 3] UpperCAmelCase = bbox[i, j, 1] UpperCAmelCase = t if bbox[i, j, 2] < bbox[i, j, 0]: UpperCAmelCase = bbox[i, j, 2] UpperCAmelCase = bbox[i, j, 0] UpperCAmelCase = t UpperCAmelCase = None if self.use_input_mask: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def _lowercase ( self : Tuple , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[Any] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = LiltModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , bbox=__lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , bbox=__lowerCamelCase , token_type_ids=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , bbox=__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Any , __lowerCamelCase : str , __lowerCamelCase : Optional[int] , ) -> List[str]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = LiltForTokenClassification(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model( __lowerCamelCase , bbox=__lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple , __lowerCamelCase : Any , __lowerCamelCase : Dict , __lowerCamelCase : List[str] , ) -> List[Any]: """simple docstring""" UpperCAmelCase = LiltForQuestionAnswering(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model( __lowerCamelCase , bbox=__lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { """input_ids""": input_ids, """bbox""": bbox, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class __lowercase ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) UpperCamelCase = ( { '''feature-extraction''': LiltModel, '''question-answering''': LiltForQuestionAnswering, '''text-classification''': LiltForSequenceClassification, '''token-classification''': LiltForTokenClassification, '''zero-shot''': LiltForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : List[str] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Dict: """simple docstring""" return True def _lowercase ( self : Dict ) -> str: """simple docstring""" UpperCAmelCase = LiltModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : int ) -> List[Any]: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase = type self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Dict: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__lowerCamelCase ) @slow def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = LiltModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @require_torch @slow class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = LiltModel.from_pretrained("""SCUT-DLVCLab/lilt-roberta-en-base""" ).to(__lowerCamelCase ) UpperCAmelCase = torch.tensor([[1, 2]] , device=__lowerCamelCase ) UpperCAmelCase = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=__lowerCamelCase ) # forward pass with torch.no_grad(): UpperCAmelCase = model(input_ids=__lowerCamelCase , bbox=__lowerCamelCase ) UpperCAmelCase = torch.Size([1, 2, 7_6_8] ) UpperCAmelCase = torch.tensor( [[-0.0_653, 0.0_950, -0.0_061], [-0.0_545, 0.0_926, -0.0_324]] , device=__lowerCamelCase , ) self.assertTrue(outputs.last_hidden_state.shape , __lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , __lowerCamelCase , atol=1e-3 ) )
627
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
1
import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .tokenization_wavaveca import WavaVecaCTCTokenizer class __lowercase ( __snake_case ): UpperCamelCase = '''Wav2Vec2FeatureExtractor''' UpperCamelCase = '''AutoTokenizer''' def __init__( self : int , __lowerCamelCase : List[str] , __lowerCamelCase : Optional[Any] ) -> List[Any]: """simple docstring""" super().__init__(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self.feature_extractor UpperCAmelCase = False @classmethod def _lowercase ( cls : List[str] , __lowerCamelCase : List[str] , **__lowerCamelCase : Dict ) -> int: """simple docstring""" try: return super().from_pretrained(__lowerCamelCase , **__lowerCamelCase ) except OSError: warnings.warn( F"""Loading a tokenizer inside {cls.__name__} from a config that does not""" """ include a `tokenizer_class` attribute is deprecated and will be """ """removed in v5. Please add `'tokenizer_class': 'Wav2Vec2CTCTokenizer'`""" """ attribute to either your `config.json` or `tokenizer_config.json` """ """file to suppress this warning: """ , __lowerCamelCase , ) UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = WavaVecaCTCTokenizer.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) return cls(feature_extractor=__lowerCamelCase , tokenizer=__lowerCamelCase ) def __call__( self : Optional[Any] , *__lowerCamelCase : Tuple , **__lowerCamelCase : str ) -> Tuple: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*__lowerCamelCase , **__lowerCamelCase ) if "raw_speech" in kwargs: warnings.warn("""Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.""" ) UpperCAmelCase = kwargs.pop("""raw_speech""" ) else: UpperCAmelCase = kwargs.pop("""audio""" , __lowerCamelCase ) UpperCAmelCase = kwargs.pop("""sampling_rate""" , __lowerCamelCase ) UpperCAmelCase = kwargs.pop("""text""" , __lowerCamelCase ) if len(__lowerCamelCase ) > 0: UpperCAmelCase = args[0] UpperCAmelCase = args[1:] if audio is None and text is None: raise ValueError("""You need to specify either an `audio` or `text` input to process.""" ) if audio is not None: UpperCAmelCase = self.feature_extractor(__lowerCamelCase , *__lowerCamelCase , sampling_rate=__lowerCamelCase , **__lowerCamelCase ) if text is not None: UpperCAmelCase = self.tokenizer(__lowerCamelCase , **__lowerCamelCase ) if text is None: return inputs elif audio is None: return encodings else: UpperCAmelCase = encodings["""input_ids"""] return inputs def _lowercase ( self : Union[str, Any] , *__lowerCamelCase : List[str] , **__lowerCamelCase : Optional[Any] ) -> Dict: """simple docstring""" if self._in_target_context_manager: return self.current_processor.pad(*__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = kwargs.pop("""input_features""" , __lowerCamelCase ) UpperCAmelCase = kwargs.pop("""labels""" , __lowerCamelCase ) if len(__lowerCamelCase ) > 0: UpperCAmelCase = args[0] UpperCAmelCase = args[1:] if input_features is not None: UpperCAmelCase = self.feature_extractor.pad(__lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) if labels is not None: UpperCAmelCase = self.tokenizer.pad(__lowerCamelCase , **__lowerCamelCase ) if labels is None: return input_features elif input_features is None: return labels else: UpperCAmelCase = labels["""input_ids"""] return input_features def _lowercase ( self : int , *__lowerCamelCase : Dict , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return self.tokenizer.batch_decode(*__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Any , *__lowerCamelCase : Tuple , **__lowerCamelCase : Dict ) -> List[Any]: """simple docstring""" return self.tokenizer.decode(*__lowerCamelCase , **__lowerCamelCase ) @contextmanager def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your audio inputs, or in a separate call.""" ) UpperCAmelCase = True UpperCAmelCase = self.tokenizer yield UpperCAmelCase = self.feature_extractor UpperCAmelCase = False
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class __lowercase ( __snake_case ): def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__lowerCamelCase , """tf_padding""" ) ) self.parent.assertTrue(hasattr(__lowerCamelCase , """depth_multiplier""" ) ) class __lowercase : def __init__( self : str , __lowerCamelCase : List[str] , __lowerCamelCase : List[str]=1_3 , __lowerCamelCase : str=3 , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : List[Any]=0.25 , __lowerCamelCase : Union[str, Any]=8 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=1_0_2_4 , __lowerCamelCase : Any=3_2 , __lowerCamelCase : Optional[int]="relu6" , __lowerCamelCase : int=0.1 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Any=True , __lowerCamelCase : int=True , __lowerCamelCase : Union[str, Any]=1_0 , __lowerCamelCase : Union[str, Any]=None , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = image_size UpperCAmelCase = depth_multiplier UpperCAmelCase = min_depth UpperCAmelCase = tf_padding UpperCAmelCase = int(last_hidden_size * depth_multiplier ) UpperCAmelCase = output_stride UpperCAmelCase = hidden_act UpperCAmelCase = classifier_dropout_prob UpperCAmelCase = use_labels UpperCAmelCase = is_training UpperCAmelCase = num_labels UpperCAmelCase = initializer_range UpperCAmelCase = scope def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCAmelCase = self.get_config() return config, pixel_values, labels, pixel_labels def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = MobileNetVaModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _lowercase ( self : List[Any] , __lowerCamelCase : str , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = MobileNetVaForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () UpperCamelCase = ( {'''feature-extraction''': MobileNetVaModel, '''image-classification''': MobileNetVaForImageClassification} if is_torch_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = MobileNetVaModelTester(self ) UpperCAmelCase = MobileNetVaConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def _lowercase ( self : List[str] ) -> Optional[int]: """simple docstring""" pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" pass def _lowercase ( self : Dict ) -> Optional[int]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Dict ) -> str: """simple docstring""" def check_hidden_states_output(__lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple ): UpperCAmelCase = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): UpperCAmelCase = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) UpperCAmelCase = outputs.hidden_states UpperCAmelCase = 2_6 self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = MobileNetVaModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->int: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(__lowerCamelCase ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""pt""" ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = torch.Size((1, 1_0_0_1) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
1
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->float: return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[list[float] | float]]: if dataset.ndim != value_array.ndim: UpperCAmelCase = ( """Wrong input data's dimensions... """ F"""dataset : {dataset.ndim}, value_array : {value_array.ndim}""" ) raise ValueError(lowerCAmelCase_ ) try: if dataset.shape[1] != value_array.shape[1]: UpperCAmelCase = ( """Wrong input data's shape... """ F"""dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}""" ) raise ValueError(lowerCAmelCase_ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("""Wrong shape""" ) if dataset.dtype != value_array.dtype: UpperCAmelCase = ( """Input data have different datatype... """ F"""dataset : {dataset.dtype}, value_array : {value_array.dtype}""" ) raise TypeError(lowerCAmelCase_ ) UpperCAmelCase = [] for value in value_array: UpperCAmelCase = euclidean(lowerCAmelCase_ , dataset[0] ) UpperCAmelCase = dataset[0].tolist() for dataset_value in dataset[1:]: UpperCAmelCase = euclidean(lowerCAmelCase_ , lowerCAmelCase_ ) if dist > temp_dist: UpperCAmelCase = temp_dist UpperCAmelCase = dataset_value.tolist() answer.append([vector, dist] ) return answer def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->float: return np.dot(lowerCAmelCase_ , lowerCAmelCase_ ) / (norm(lowerCAmelCase_ ) * norm(lowerCAmelCase_ )) if __name__ == "__main__": import doctest doctest.testmod()
627
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
1
def _UpperCamelCase ( lowerCAmelCase_ ) ->str: # noqa: E741 UpperCAmelCase = len(lowerCAmelCase_ ) UpperCAmelCase = 0 UpperCAmelCase = [0] * n UpperCAmelCase = [False] * n UpperCAmelCase = [False] * n def dfs(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): if parent == root: out_edge_count += 1 UpperCAmelCase = True UpperCAmelCase = at for to in l[at]: if to == parent: pass elif not visited[to]: UpperCAmelCase = dfs(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: UpperCAmelCase = True # AP found via cycle if at == low[to]: UpperCAmelCase = True else: UpperCAmelCase = min(low[at] , lowerCAmelCase_ ) return out_edge_count for i in range(lowerCAmelCase_ ): if not visited[i]: UpperCAmelCase = 0 UpperCAmelCase = dfs(lowerCAmelCase_ , lowerCAmelCase_ , -1 , lowerCAmelCase_ ) UpperCAmelCase = out_edge_count > 1 for x in range(len(lowerCAmelCase_ ) ): if is_art[x] is True: print(lowerCAmelCase_ ) # Adjacency list of graph __a = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
1
import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __a = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. __a = importlib.util.spec_from_file_location( """transformers""", os.path.join(PATH_TO_TRANSFORMERS, """__init__.py"""), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) __a = spec.loader.load_module() __a = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __a = re.compile("""\[(.+?)\]\((https://huggingface\.co/.+?)\)""") __a = { """CLIPConfigMixin""", """DecisionTransformerConfigMixin""", """EncoderDecoderConfigMixin""", """RagConfigMixin""", """SpeechEncoderDecoderConfigMixin""", """VisionEncoderDecoderConfigMixin""", """VisionTextDualEncoderConfigMixin""", } def _UpperCamelCase ( ) ->int: UpperCAmelCase = [] for config_class in list(CONFIG_MAPPING.values() ): UpperCAmelCase = False # source code of `config_class` UpperCAmelCase = inspect.getsource(lowerCAmelCase_ ) UpperCAmelCase = _re_checkpoint.findall(lowerCAmelCase_ ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` UpperCAmelCase , UpperCAmelCase = checkpoint # verify the checkpoint name corresponds to the checkpoint link UpperCAmelCase = F"""https://huggingface.co/{ckpt_name}""" if ckpt_link == ckpt_link_from_name: UpperCAmelCase = True break UpperCAmelCase = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: UpperCAmelCase = """\n""".join(sorted(lowerCAmelCase_ ) ) raise ValueError(F"""The following configurations don't contain any valid checkpoint:\n{message}""" ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
627
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
1
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
1
import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase = FlaxDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=__lowerCamelCase , cache_dir=__lowerCamelCase ) UpperCAmelCase = [t[-1] for t in os.walk(os.path.join(__lowerCamelCase , os.listdir(__lowerCamelCase )[0] , """snapshots""" ) )] UpperCAmelCase = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(""".bin""" ) for f in files ) @slow @require_flax class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=__lowerCamelCase ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 4 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 6_4, 6_4, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_514_745 ) < 1e-3 assert np.abs(np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 49_947.875 ) < 5e-1 UpperCAmelCase = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(__lowerCamelCase ) == num_samples def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""flax""" , safety_checker=__lowerCamelCase ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_652_401) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_383_808.2) ) < 5e-1 def _lowercase ( self : int ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCamelCase ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_003_906) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_373_516.75) ) < 5e-1 def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa ) UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_003_906) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_373_516.75) ) < 5e-1 def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = FlaxDDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , set_alpha_to_one=__lowerCamelCase , steps_offset=1 , ) UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , scheduler=__lowerCamelCase , safety_checker=__lowerCamelCase , ) UpperCAmelCase = scheduler.create_state() UpperCAmelCase = scheduler_state UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = 5_0 UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) # shard inputs and rng UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = jax.random.split(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.045_043_945) ) < 1e-3 assert np.abs((np.abs(__lowerCamelCase , dtype=np.floataa ).sum() - 2_347_693.5) ) < 5e-1 def _lowercase ( self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = jax.random.split(jax.random.PRNGKey(0 ) , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCamelCase , ) UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) UpperCAmelCase = images[2, 0, 2_5_6, 1_0:1_7, 1] # With memory efficient attention UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCamelCase , use_memory_efficient_attention=__lowerCamelCase , ) UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = pipeline.prepare_inputs(__lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = pipeline(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , jit=__lowerCamelCase ).images assert images_eff.shape == (num_samples, 1, 5_1_2, 5_1_2, 3) UpperCAmelCase = images[2, 0, 2_5_6, 1_0:1_7, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1e-2
627
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__snake_case ) , '''Tatoeba directory does not exist.''' ) class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() return TatoebaConverter(save_dir=__lowerCamelCase ) @slow def _lowercase ( self : int ) -> str: """simple docstring""" self.resolver.convert_models(["""heb-eng"""] ) @slow def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.resolver.write_model_card("""opus-mt-he-en""" , dry_run=__lowerCamelCase ) assert mmeta["long_pair"] == "heb-eng"
627
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
1
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __a = logging.get_logger(__name__) @add_end_docstrings(__snake_case ) class __lowercase ( __snake_case ): def __init__( self : Union[str, Any] , *__lowerCamelCase : Optional[Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" super().__init__(*__lowerCamelCase , **__lowerCamelCase ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _lowercase ( self : Optional[int] , __lowerCamelCase : Optional[int]=None ) -> Dict: """simple docstring""" UpperCAmelCase = {} if top_k is not None: UpperCAmelCase = top_k return {}, {}, postprocess_params def __call__( self : Optional[int] , __lowerCamelCase : Union[str, List[str], "Image.Image", List["Image.Image"]] , **__lowerCamelCase : Optional[int] ) -> int: """simple docstring""" return super().__call__(__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : List[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = load_image(__lowerCamelCase ) UpperCAmelCase = self.image_processor(images=__lowerCamelCase , return_tensors=self.framework ) return model_inputs def _lowercase ( self : List[str] , __lowerCamelCase : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.model(**__lowerCamelCase ) return model_outputs def _lowercase ( self : Dict , __lowerCamelCase : Optional[Any] , __lowerCamelCase : str=5 ) -> Any: """simple docstring""" if top_k > self.model.config.num_labels: UpperCAmelCase = self.model.config.num_labels if self.framework == "pt": UpperCAmelCase = model_outputs.logits.softmax(-1 )[0] UpperCAmelCase , UpperCAmelCase = probs.topk(__lowerCamelCase ) elif self.framework == "tf": UpperCAmelCase = stable_softmax(model_outputs.logits , axis=-1 )[0] UpperCAmelCase = tf.math.top_k(__lowerCamelCase , k=__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(F"""Unsupported framework: {self.framework}""" ) UpperCAmelCase = scores.tolist() UpperCAmelCase = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(__lowerCamelCase , __lowerCamelCase )]
627
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
1
import sys __a = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043557""" """66896648950445244523161731856403098711121722383113""" """62229893423380308135336276614282806444486645238749""" """30358907296290491560440772390713810515859307960866""" """70172427121883998797908792274921901699720888093776""" """65727333001053367881220235421809751254540594752243""" """52584907711670556013604839586446706324415722155397""" """53697817977846174064955149290862569321978468622482""" """83972241375657056057490261407972968652414535100474""" """82166370484403199890008895243450658541227588666881""" """16427171479924442928230863465674813919123162824586""" """17866458359124566529476545682848912883142607690042""" """24219022671055626321111109370544217506941658960408""" """07198403850962455444362981230987879927244284909188""" """84580156166097919133875499200524063689912560717606""" """05886116467109405077541002256983155200055935729725""" """71636269561882670428252483600823257530420752963450""" ) def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = 1 for digit in s: product *= int(lowerCAmelCase_ ) return product def _UpperCamelCase ( lowerCAmelCase_ = N ) ->int: UpperCAmelCase = -sys.maxsize - 1 UpperCAmelCase = n[:1_3] UpperCAmelCase = 1_3 while cur_index < len(lowerCAmelCase_ ) - 1_3: if int(n[cur_index] ) >= int(substr[0] ): UpperCAmelCase = substr[1:] + n[cur_index] cur_index += 1 else: UpperCAmelCase = max(lowerCAmelCase_ , str_eval(lowerCAmelCase_ ) ) UpperCAmelCase = n[cur_index : cur_index + 1_3] cur_index += 1_3 return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
627
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
1
from collections.abc import Callable class __lowercase : def __init__( self : List[str] , __lowerCamelCase : Callable | None = None ) -> None: """simple docstring""" UpperCAmelCase = [] # Stores indexes of each item for supporting updates and deletion. UpperCAmelCase = {} # Stores current size of heap. UpperCAmelCase = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. UpperCAmelCase = key or (lambda __lowerCamelCase : x) def _lowercase ( self : List[Any] , __lowerCamelCase : int ) -> int | None: """simple docstring""" return int((i - 1) / 2 ) if i > 0 else None def _lowercase ( self : List[str] , __lowerCamelCase : int ) -> int | None: """simple docstring""" UpperCAmelCase = int(2 * i + 1 ) return left if 0 < left < self.size else None def _lowercase ( self : Dict , __lowerCamelCase : int ) -> int | None: """simple docstring""" UpperCAmelCase = int(2 * i + 2 ) return right if 0 < right < self.size else None def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : int ) -> None: """simple docstring""" UpperCAmelCase , UpperCAmelCase = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. UpperCAmelCase , UpperCAmelCase = self.arr[j], self.arr[i] def _lowercase ( self : int , __lowerCamelCase : int , __lowerCamelCase : int ) -> bool: """simple docstring""" return self.arr[i][1] < self.arr[j][1] def _lowercase ( self : Optional[Any] , __lowerCamelCase : int ) -> int: """simple docstring""" UpperCAmelCase = self._left(__lowerCamelCase ) UpperCAmelCase = self._right(__lowerCamelCase ) UpperCAmelCase = i if left is not None and not self._cmp(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = left if right is not None and not self._cmp(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = right return valid_parent def _lowercase ( self : Optional[Any] , __lowerCamelCase : int ) -> None: """simple docstring""" UpperCAmelCase = self._parent(__lowerCamelCase ) while parent is not None and not self._cmp(__lowerCamelCase , __lowerCamelCase ): self._swap(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = parent, self._parent(__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : int ) -> None: """simple docstring""" UpperCAmelCase = self._get_valid_parent(__lowerCamelCase ) while valid_parent != index: self._swap(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = valid_parent, self._get_valid_parent(__lowerCamelCase ) def _lowercase ( self : int , __lowerCamelCase : int , __lowerCamelCase : int ) -> None: """simple docstring""" if item not in self.pos_map: return UpperCAmelCase = self.pos_map[item] UpperCAmelCase = [item, self.key(__lowerCamelCase )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(__lowerCamelCase ) self._heapify_down(__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : int ) -> None: """simple docstring""" if item not in self.pos_map: return UpperCAmelCase = self.pos_map[item] del self.pos_map[item] UpperCAmelCase = self.arr[self.size - 1] UpperCAmelCase = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(__lowerCamelCase ) self._heapify_down(__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : int , __lowerCamelCase : int ) -> None: """simple docstring""" UpperCAmelCase = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(__lowerCamelCase )] ) else: UpperCAmelCase = [item, self.key(__lowerCamelCase )] UpperCAmelCase = self.size self.size += 1 self._heapify_up(self.size - 1 ) def _lowercase ( self : Optional[Any] ) -> tuple | None: """simple docstring""" return self.arr[0] if self.size else None def _lowercase ( self : Dict ) -> tuple | None: """simple docstring""" UpperCAmelCase = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def _UpperCamelCase ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
627
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
1
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
1
def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0 ) ->int: return sum(e for e in range(3 , lowerCAmelCase_ ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F"""{solution() = }""")
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
1
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __lowercase ( __snake_case ): UpperCamelCase = ['''image_processor''', '''tokenizer'''] UpperCamelCase = '''AutoImageProcessor''' UpperCamelCase = '''AutoTokenizer''' def __init__( self : Optional[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ) -> List[str]: """simple docstring""" super().__init__(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self.image_processor def __call__( self : List[Any] , __lowerCamelCase : Any=None , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=None , **__lowerCamelCase : int ) -> Any: """simple docstring""" if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if images is not None: UpperCAmelCase = self.image_processor(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__lowerCamelCase ) , tensor_type=__lowerCamelCase ) def _lowercase ( self : Any , *__lowerCamelCase : List[Any] , **__lowerCamelCase : Tuple ) -> Tuple: """simple docstring""" return self.tokenizer.batch_decode(*__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : int , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : List[str] ) -> Any: """simple docstring""" return self.tokenizer.decode(*__lowerCamelCase , **__lowerCamelCase ) @property def _lowercase ( self : List[str] ) -> Dict: """simple docstring""" return ["input_ids", "attention_mask", "pixel_values"]
627
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
1
from collections.abc import Sequence from queue import Queue class __lowercase : def __init__( self : Dict , __lowerCamelCase : Dict , __lowerCamelCase : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int=None , __lowerCamelCase : Optional[int]=None ) -> Dict: """simple docstring""" UpperCAmelCase = start UpperCAmelCase = end UpperCAmelCase = val UpperCAmelCase = (start + end) // 2 UpperCAmelCase = left UpperCAmelCase = right def __repr__( self : int ) -> Any: """simple docstring""" return F"""SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})""" class __lowercase : def __init__( self : str , __lowerCamelCase : Sequence , __lowerCamelCase : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = collection UpperCAmelCase = function if self.collection: UpperCAmelCase = self._build_tree(0 , len(__lowerCamelCase ) - 1 ) def _lowercase ( self : Any , __lowerCamelCase : Tuple , __lowerCamelCase : Dict ) -> Union[str, Any]: """simple docstring""" self._update_tree(self.root , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Any ) -> Tuple: """simple docstring""" return self._query_range(self.root , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : int , __lowerCamelCase : str , __lowerCamelCase : List[Any] ) -> List[str]: """simple docstring""" if start == end: return SegmentTreeNode(__lowerCamelCase , __lowerCamelCase , self.collection[start] ) UpperCAmelCase = (start + end) // 2 UpperCAmelCase = self._build_tree(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._build_tree(mid + 1 , __lowerCamelCase ) return SegmentTreeNode(__lowerCamelCase , __lowerCamelCase , self.fn(left.val , right.val ) , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Any , __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple ) -> Optional[Any]: """simple docstring""" if node.start == i and node.end == i: UpperCAmelCase = val return if i <= node.mid: self._update_tree(node.left , __lowerCamelCase , __lowerCamelCase ) else: self._update_tree(node.right , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self.fn(node.left.val , node.right.val ) def _lowercase ( self : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : List[str] , __lowerCamelCase : Optional[int] ) -> Dict: """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , __lowerCamelCase , __lowerCamelCase ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , __lowerCamelCase , node.mid ) , self._query_range(node.right , node.mid + 1 , __lowerCamelCase ) , ) else: # range in right child tree return self._query_range(node.right , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Dict ) -> Any: """simple docstring""" if self.root is not None: UpperCAmelCase = Queue() queue.put(self.root ) while not queue.empty(): UpperCAmelCase = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print("""*""" * 50) __a = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class __lowercase ( unittest.TestCase ): def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = 1 UpperCAmelCase = 3 UpperCAmelCase = (3_2, 3_2) UpperCAmelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(__lowerCamelCase ) return image @property def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , ) return model @property def _lowercase ( self : int ) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = RobertaSeriesConfig( hidden_size=3_2 , project_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_0_0_6 , ) return RobertaSeriesModelWithTransformation(__lowerCamelCase ) @property def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" def extract(*__lowerCamelCase : Dict , **__lowerCamelCase : Dict ): class __lowercase : def __init__( self : Optional[Any] ) -> str: """simple docstring""" UpperCAmelCase = torch.ones([0] ) def _lowercase ( self : str , __lowerCamelCase : List[Any] ) -> Tuple: """simple docstring""" self.pixel_values.to(__lowerCamelCase ) return self return Out() return extract def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.dummy_cond_unet UpperCAmelCase = PNDMScheduler(skip_prk_steps=__lowerCamelCase ) UpperCAmelCase = self.dummy_vae UpperCAmelCase = self.dummy_text_encoder UpperCAmelCase = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""" ) UpperCAmelCase = 7_7 UpperCAmelCase = self.dummy_image.to(__lowerCamelCase ) UpperCAmelCase = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk UpperCAmelCase = AltDiffusionImgaImgPipeline( unet=__lowerCamelCase , scheduler=__lowerCamelCase , vae=__lowerCamelCase , text_encoder=__lowerCamelCase , tokenizer=__lowerCamelCase , safety_checker=__lowerCamelCase , feature_extractor=self.dummy_extractor , ) UpperCAmelCase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=__lowerCamelCase ) UpperCAmelCase = alt_pipe.to(__lowerCamelCase ) alt_pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = """A painting of a squirrel eating a burger""" UpperCAmelCase = torch.Generator(device=__lowerCamelCase ).manual_seed(0 ) UpperCAmelCase = alt_pipe( [prompt] , generator=__lowerCamelCase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , image=__lowerCamelCase , ) UpperCAmelCase = output.images UpperCAmelCase = torch.Generator(device=__lowerCamelCase ).manual_seed(0 ) UpperCAmelCase = alt_pipe( [prompt] , generator=__lowerCamelCase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , image=__lowerCamelCase , return_dict=__lowerCamelCase , )[0] UpperCAmelCase = image[0, -3:, -3:, -1] UpperCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) UpperCAmelCase = np.array([0.4_427, 0.3_731, 0.4_249, 0.4_941, 0.4_546, 0.4_148, 0.4_193, 0.4_666, 0.4_499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5e-3 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" UpperCAmelCase = self.dummy_cond_unet UpperCAmelCase = PNDMScheduler(skip_prk_steps=__lowerCamelCase ) UpperCAmelCase = self.dummy_vae UpperCAmelCase = self.dummy_text_encoder UpperCAmelCase = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""" ) UpperCAmelCase = 7_7 UpperCAmelCase = self.dummy_image.to(__lowerCamelCase ) # put models in fp16 UpperCAmelCase = unet.half() UpperCAmelCase = vae.half() UpperCAmelCase = bert.half() # make sure here that pndm scheduler skips prk UpperCAmelCase = AltDiffusionImgaImgPipeline( unet=__lowerCamelCase , scheduler=__lowerCamelCase , vae=__lowerCamelCase , text_encoder=__lowerCamelCase , tokenizer=__lowerCamelCase , safety_checker=__lowerCamelCase , feature_extractor=self.dummy_extractor , ) UpperCAmelCase = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=__lowerCamelCase ) UpperCAmelCase = alt_pipe.to(__lowerCamelCase ) alt_pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = """A painting of a squirrel eating a burger""" UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = alt_pipe( [prompt] , generator=__lowerCamelCase , num_inference_steps=2 , output_type="""np""" , image=__lowerCamelCase , ).images assert image.shape == (1, 3_2, 3_2, 3) @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def _lowercase ( self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) # resize to resolution that is divisible by 8 but not 16 or 32 UpperCAmelCase = init_image.resize((7_6_0, 5_0_4) ) UpperCAmelCase = """BAAI/AltDiffusion""" UpperCAmelCase = AltDiffusionImgaImgPipeline.from_pretrained( __lowerCamelCase , safety_checker=__lowerCamelCase , ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) pipe.enable_attention_slicing() UpperCAmelCase = """A fantasy landscape, trending on artstation""" UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = pipe( prompt=__lowerCamelCase , image=__lowerCamelCase , strength=0.75 , guidance_scale=7.5 , generator=__lowerCamelCase , output_type="""np""" , ) UpperCAmelCase = output.images[0] UpperCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 7_6_0, 3) UpperCAmelCase = np.array([0.9_358, 0.9_397, 0.9_599, 0.9_901, 1.0_000, 1.0_000, 0.9_882, 1.0_000, 1.0_000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : str ) -> Tuple: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) UpperCAmelCase = init_image.resize((7_6_8, 5_1_2) ) UpperCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy""" ) UpperCAmelCase = """BAAI/AltDiffusion""" UpperCAmelCase = AltDiffusionImgaImgPipeline.from_pretrained( __lowerCamelCase , safety_checker=__lowerCamelCase , ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) pipe.enable_attention_slicing() UpperCAmelCase = """A fantasy landscape, trending on artstation""" UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = pipe( prompt=__lowerCamelCase , image=__lowerCamelCase , strength=0.75 , guidance_scale=7.5 , generator=__lowerCamelCase , output_type="""np""" , ) UpperCAmelCase = output.images[0] assert image.shape == (5_1_2, 7_6_8, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1e-2
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets __a = """\ @inproceedings{snover-etal-2006-study, title = \"A Study of Translation Edit Rate with Targeted Human Annotation\", author = \"Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John\", booktitle = \"Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers\", month = aug # \" 8-12\", year = \"2006\", address = \"Cambridge, Massachusetts, USA\", publisher = \"Association for Machine Translation in the Americas\", url = \"https://aclanthology.org/2006.amta-papers.25\", pages = \"223--231\", } @inproceedings{post-2018-call, title = \"A Call for Clarity in Reporting {BLEU} Scores\", author = \"Post, Matt\", booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\", month = oct, year = \"2018\", address = \"Belgium, Brussels\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W18-6319\", pages = \"186--191\", } """ __a = """\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. """ __a = """ Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: 'score' (float): TER score (num_edits / sum_ref_lengths * 100) 'num_edits' (int): The cumulative number of edits 'ref_length' (float): The cumulative average reference length Examples: Example 1: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 150.0, 'num_edits': 15, 'ref_length': 10.0} Example 2: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 62.5, 'num_edits': 5, 'ref_length': 8.0} Example 3: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {'score': 57.14285714285714, 'num_edits': 6, 'ref_length': 10.5} Example 4: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 0.0, 'num_edits': 0, 'ref_length': 8.0} Example 5: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 100.0, 'num_edits': 10, 'ref_length': 10.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""http://www.cs.umd.edu/~snover/tercom/""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#ter"""] , reference_urls=[ """https://github.com/jhclark/tercom""", ] , ) def _lowercase ( self : List[str] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[str] , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , ) -> Tuple: """simple docstring""" UpperCAmelCase = len(references[0] ) if any(len(__lowerCamelCase ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) UpperCAmelCase = [[refs[i] for refs in references] for i in range(__lowerCamelCase )] UpperCAmelCase = TER( normalized=__lowerCamelCase , no_punct=__lowerCamelCase , asian_support=__lowerCamelCase , case_sensitive=__lowerCamelCase , ) UpperCAmelCase = sb_ter.corpus_score(__lowerCamelCase , __lowerCamelCase ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
627
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __a = logging.get_logger(__name__) __a = {"""vocab_file""": """spm_char.model"""} __a = { """vocab_file""": { """microsoft/speecht5_asr""": """https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model""", """microsoft/speecht5_tts""": """https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model""", """microsoft/speecht5_vc""": """https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model""", } } __a = { """microsoft/speecht5_asr""": 1024, """microsoft/speecht5_tts""": 1024, """microsoft/speecht5_vc""": 1024, } class __lowercase ( __snake_case ): UpperCamelCase = VOCAB_FILES_NAMES UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase = ['''input_ids''', '''attention_mask'''] def __init__( self : Optional[Any] , __lowerCamelCase : Any , __lowerCamelCase : List[str]="<s>" , __lowerCamelCase : Tuple="</s>" , __lowerCamelCase : Any="<unk>" , __lowerCamelCase : Union[str, Any]="<pad>" , __lowerCamelCase : Optional[Dict[str, Any]] = None , **__lowerCamelCase : Any , ) -> None: """simple docstring""" UpperCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , unk_token=__lowerCamelCase , pad_token=__lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__lowerCamelCase , ) UpperCAmelCase = vocab_file UpperCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__lowerCamelCase ) @property def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" return self.sp_model.get_piece_size() def _lowercase ( self : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = {self.convert_ids_to_tokens(__lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.__dict__.copy() UpperCAmelCase = None return state def __setstate__( self : int , __lowerCamelCase : List[str] ) -> List[Any]: """simple docstring""" UpperCAmelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCAmelCase = {} UpperCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self : str , __lowerCamelCase : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(__lowerCamelCase , out_type=__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : Dict ) -> Tuple: """simple docstring""" return self.sp_model.piece_to_id(__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.sp_model.IdToPiece(__lowerCamelCase ) return token def _lowercase ( self : Optional[int] , __lowerCamelCase : int ) -> List[str]: """simple docstring""" UpperCAmelCase = [] UpperCAmelCase = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__lowerCamelCase ) + token UpperCAmelCase = [] else: current_sub_tokens.append(__lowerCamelCase ) out_string += self.sp_model.decode(__lowerCamelCase ) return out_string.strip() def _lowercase ( self : int , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[Any]=None ) -> List[int]: """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _lowercase ( self : Dict , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None , __lowerCamelCase : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase ) UpperCAmelCase = [1] if token_ids_a is None: return ([0] * len(__lowerCamelCase )) + suffix_ones return ([0] * len(__lowerCamelCase )) + ([0] * len(__lowerCamelCase )) + suffix_ones def _lowercase ( self : Tuple , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__lowerCamelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase = os.path.join( __lowerCamelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__lowerCamelCase , """wb""" ) as fi: UpperCAmelCase = self.sp_model.serialized_model_proto() fi.write(__lowerCamelCase ) return (out_vocab_file,)
627
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
1
import unittest import numpy as np from transformers import MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING, TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING from transformers.pipelines import AudioClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_torchaudio, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class __lowercase ( unittest.TestCase ): UpperCamelCase = MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING UpperCamelCase = TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING def _lowercase ( self : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = AudioClassificationPipeline(model=__lowerCamelCase , feature_extractor=__lowerCamelCase ) # test with a raw waveform UpperCAmelCase = np.zeros((3_4_0_0_0,) ) UpperCAmelCase = np.zeros((1_4_0_0_0,) ) return audio_classifier, [audioa, audio] def _lowercase ( self : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int ) -> int: """simple docstring""" UpperCAmelCase , UpperCAmelCase = examples UpperCAmelCase = audio_classifier(__lowerCamelCase ) # by default a model is initialized with num_labels=2 self.assertEqual( __lowerCamelCase , [ {"""score""": ANY(__lowerCamelCase ), """label""": ANY(__lowerCamelCase )}, {"""score""": ANY(__lowerCamelCase ), """label""": ANY(__lowerCamelCase )}, ] , ) UpperCAmelCase = audio_classifier(__lowerCamelCase , top_k=1 ) self.assertEqual( __lowerCamelCase , [ {"""score""": ANY(__lowerCamelCase ), """label""": ANY(__lowerCamelCase )}, ] , ) self.run_torchaudio(__lowerCamelCase ) @require_torchaudio def _lowercase ( self : str , __lowerCamelCase : int ) -> Any: """simple docstring""" import datasets # test with a local file UpperCAmelCase = datasets.load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) UpperCAmelCase = dataset[0]["""audio"""]["""array"""] UpperCAmelCase = audio_classifier(__lowerCamelCase ) self.assertEqual( __lowerCamelCase , [ {"""score""": ANY(__lowerCamelCase ), """label""": ANY(__lowerCamelCase )}, {"""score""": ANY(__lowerCamelCase ), """label""": ANY(__lowerCamelCase )}, ] , ) @require_torch def _lowercase ( self : List[str] ) -> List[str]: """simple docstring""" UpperCAmelCase = """anton-l/wav2vec2-random-tiny-classifier""" UpperCAmelCase = pipeline("""audio-classification""" , model=__lowerCamelCase ) UpperCAmelCase = np.ones((8_0_0_0,) ) UpperCAmelCase = audio_classifier(__lowerCamelCase , top_k=4 ) UpperCAmelCase = [ {"""score""": 0.0_842, """label""": """no"""}, {"""score""": 0.0_838, """label""": """up"""}, {"""score""": 0.0_837, """label""": """go"""}, {"""score""": 0.0_834, """label""": """right"""}, ] UpperCAmelCase = [ {"""score""": 0.0_845, """label""": """stop"""}, {"""score""": 0.0_844, """label""": """on"""}, {"""score""": 0.0_841, """label""": """right"""}, {"""score""": 0.0_834, """label""": """left"""}, ] self.assertIn(nested_simplify(__lowerCamelCase , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) UpperCAmelCase = {"""array""": np.ones((8_0_0_0,) ), """sampling_rate""": audio_classifier.feature_extractor.sampling_rate} UpperCAmelCase = audio_classifier(__lowerCamelCase , top_k=4 ) self.assertIn(nested_simplify(__lowerCamelCase , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) @require_torch @slow def _lowercase ( self : List[str] ) -> Optional[int]: """simple docstring""" import datasets UpperCAmelCase = """superb/wav2vec2-base-superb-ks""" UpperCAmelCase = pipeline("""audio-classification""" , model=__lowerCamelCase ) UpperCAmelCase = datasets.load_dataset("""anton-l/superb_dummy""" , """ks""" , split="""test""" ) UpperCAmelCase = np.array(dataset[3]["""speech"""] , dtype=np.floataa ) UpperCAmelCase = audio_classifier(__lowerCamelCase , top_k=4 ) self.assertEqual( nested_simplify(__lowerCamelCase , decimals=3 ) , [ {"""score""": 0.981, """label""": """go"""}, {"""score""": 0.007, """label""": """up"""}, {"""score""": 0.006, """label""": """_unknown_"""}, {"""score""": 0.001, """label""": """down"""}, ] , ) @require_tf @unittest.skip("""Audio classification is not implemented for TF""" ) def _lowercase ( self : Any ) -> Any: """simple docstring""" pass
627
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
1
def _UpperCamelCase ( lowerCAmelCase_ ) ->list[list]: UpperCAmelCase = current_set.copy() for row_index, row in enumerate(lowerCAmelCase_ ): UpperCAmelCase = row[0] for column_index, column in enumerate(lowerCAmelCase_ ): if magnitude == 0: UpperCAmelCase = column continue UpperCAmelCase = column / magnitude # Subtract to cancel term UpperCAmelCase = current_set[0] UpperCAmelCase = [first_row] UpperCAmelCase = current_set[1::] for row in current_set: UpperCAmelCase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(lowerCAmelCase_ ) continue for column_index in range(len(lowerCAmelCase_ ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(lowerCAmelCase_ ) # Create next recursion iteration set if len(final_set[0] ) != 3: UpperCAmelCase = final_set[0] UpperCAmelCase = [] UpperCAmelCase = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) UpperCAmelCase = simplify(lowerCAmelCase_ ) for i in range(len(lowerCAmelCase_ ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , lowerCAmelCase_ ) UpperCAmelCase = resultant return final_set def _UpperCamelCase ( lowerCAmelCase_ ) ->list: if len(lowerCAmelCase_ ) == 0: raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) UpperCAmelCase = len(lowerCAmelCase_ ) + 1 if any(len(lowerCAmelCase_ ) != _length for item in equations ): raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) for row in equations: if any(not isinstance(lowerCAmelCase_ , (int, float) ) for column in row ): raise ValueError("""solve_simultaneous() requires lists of integers""" ) if len(lowerCAmelCase_ ) == 1: return [equations[0][-1] / equations[0][0]] UpperCAmelCase = equations.copy() if any(0 in row for row in data_set ): UpperCAmelCase = data_set.copy() UpperCAmelCase = [] for row_index, row in enumerate(lowerCAmelCase_ ): if 0 not in row: UpperCAmelCase = data_set.pop(lowerCAmelCase_ ) break if not full_row: raise ValueError("""solve_simultaneous() requires at least 1 full equation""" ) data_set.insert(0 , lowerCAmelCase_ ) UpperCAmelCase = data_set.copy() UpperCAmelCase = simplify(lowerCAmelCase_ ) UpperCAmelCase = simplified[::-1] UpperCAmelCase = [] for row in simplified: UpperCAmelCase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue UpperCAmelCase = row.copy()[: len(lowerCAmelCase_ ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(lowerCAmelCase_ ) == 0: solutions.append(0 ) continue UpperCAmelCase = temp_row[1::] UpperCAmelCase = temp_row[::-1] for column_index, column in enumerate(lowerCAmelCase_ ): current_solution -= column * solutions[column_index] solutions.append(lowerCAmelCase_ ) UpperCAmelCase = [] for item in solutions: final.append(float(round(lowerCAmelCase_ , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() __a = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
627
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
1
from math import isqrt, loga def _UpperCamelCase ( lowerCAmelCase_ ) ->list[int]: UpperCAmelCase = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = False return [i for i in range(2 , lowerCAmelCase_ ) if is_prime[i]] def _UpperCamelCase ( lowerCAmelCase_ = 8_0_0_8_0_0 , lowerCAmelCase_ = 8_0_0_8_0_0 ) ->int: UpperCAmelCase = degree * loga(lowerCAmelCase_ ) UpperCAmelCase = int(lowerCAmelCase_ ) UpperCAmelCase = calculate_prime_numbers(lowerCAmelCase_ ) UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = len(lowerCAmelCase_ ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(F"""{solution() = }""")
627
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Tuple=7 , __lowerCamelCase : int=3 , __lowerCamelCase : List[str]=1_0 , __lowerCamelCase : Any=1_8 , __lowerCamelCase : str=3_0 , __lowerCamelCase : Dict=4_0_0 , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Any=None , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=[0.5, 0.5, 0.5] , __lowerCamelCase : Any=[0.5, 0.5, 0.5] , __lowerCamelCase : str=None , ) -> List[Any]: """simple docstring""" UpperCAmelCase = size if size is not None else {"""shortest_edge""": 1_8} UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 1_8, """width""": 1_8} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = num_frames UpperCAmelCase = image_size UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_normalize UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = crop_size def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = VivitImageProcessor if is_vision_available() else None def _lowercase ( self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase = VivitImageProcessingTester(self ) @property def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_center_crop""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """size""" ) ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 1_8} ) self.assertEqual(image_processor.crop_size , {"""height""": 1_8, """width""": 1_8} ) UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 4_2} ) self.assertEqual(image_processor.crop_size , {"""height""": 8_4, """width""": 8_4} ) def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos UpperCAmelCase = prepare_video_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase ) for video in video_inputs: self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _lowercase ( self : List[str] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_video_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , numpify=__lowerCamelCase ) for video in video_inputs: self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _lowercase ( self : List[Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_video_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , torchify=__lowerCamelCase ) for video in video_inputs: self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(video_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
1
import requests __a = """YOUR API KEY""" def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = giphy_api_key ) ->list: UpperCAmelCase = """+""".join(query.split() ) UpperCAmelCase = F"""https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}""" UpperCAmelCase = requests.get(lowerCAmelCase_ ).json()["""data"""] return [gif["url"] for gif in gifs] if __name__ == "__main__": print("""\n""".join(get_gifs("""space ship""")))
627
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __lowercase ( __snake_case ): UpperCamelCase = ['''image_processor''', '''tokenizer'''] UpperCamelCase = '''CLIPImageProcessor''' UpperCamelCase = ('''XLMRobertaTokenizer''', '''XLMRobertaTokenizerFast''') def __init__( self : List[Any] , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Dict=None , **__lowerCamelCase : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __lowerCamelCase , ) UpperCAmelCase = kwargs.pop("""feature_extractor""" ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__lowerCamelCase , __lowerCamelCase ) def __call__( self : Tuple , __lowerCamelCase : List[str]=None , __lowerCamelCase : Tuple=None , __lowerCamelCase : Dict=None , **__lowerCamelCase : Union[str, Any] ) -> Dict: """simple docstring""" if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase = self.tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if images is not None: UpperCAmelCase = self.image_processor(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__lowerCamelCase ) , tensor_type=__lowerCamelCase ) def _lowercase ( self : List[str] , *__lowerCamelCase : Tuple , **__lowerCamelCase : Dict ) -> Optional[Any]: """simple docstring""" return self.tokenizer.batch_decode(*__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Any , *__lowerCamelCase : str , **__lowerCamelCase : List[Any] ) -> int: """simple docstring""" return self.tokenizer.decode(*__lowerCamelCase , **__lowerCamelCase ) @property def _lowercase ( self : int ) -> Tuple: """simple docstring""" UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
627
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
1
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = DDIMPipeline UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_PARAMS UpperCamelCase = PipelineTesterMixin.required_optional_params - { '''num_images_per_prompt''', '''latents''', '''callback''', '''callback_steps''', } UpperCamelCase = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS UpperCamelCase = False def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) UpperCAmelCase = DDIMScheduler() UpperCAmelCase = {"""unet""": unet, """scheduler""": scheduler} return components def _lowercase ( self : int , __lowerCamelCase : Dict , __lowerCamelCase : Optional[Any]=0 ) -> Optional[Any]: """simple docstring""" if str(__lowerCamelCase ).startswith("""mps""" ): UpperCAmelCase = torch.manual_seed(__lowerCamelCase ) else: UpperCAmelCase = torch.Generator(device=__lowerCamelCase ).manual_seed(__lowerCamelCase ) UpperCAmelCase = { """batch_size""": 1, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = """cpu""" UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**__lowerCamelCase ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = self.get_dummy_inputs(__lowerCamelCase ) UpperCAmelCase = pipe(**__lowerCamelCase ).images UpperCAmelCase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 3_2, 3_2, 3) ) UpperCAmelCase = np.array( [1.0_0_0e0_0, 5.7_1_7e-0_1, 4.7_1_7e-0_1, 1.0_0_0e0_0, 0.0_0_0e0_0, 1.0_0_0e0_0, 3.0_0_0e-0_4, 0.0_0_0e0_0, 9.0_0_0e-0_4] ) UpperCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__lowerCamelCase , 1e-3 ) def _lowercase ( self : Dict ) -> Optional[int]: """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" super().test_save_load_local(expected_max_difference=3e-3 ) def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3e-3 ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = """google/ddpm-cifar10-32""" UpperCAmelCase = UNetaDModel.from_pretrained(__lowerCamelCase ) UpperCAmelCase = DDIMScheduler() UpperCAmelCase = DDIMPipeline(unet=__lowerCamelCase , scheduler=__lowerCamelCase ) ddim.to(__lowerCamelCase ) ddim.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = ddim(generator=__lowerCamelCase , eta=0.0 , output_type="""numpy""" ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) UpperCAmelCase = np.array([0.1_723, 0.1_617, 0.1_600, 0.1_626, 0.1_497, 0.1_513, 0.1_505, 0.1_442, 0.1_453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" UpperCAmelCase = """google/ddpm-ema-bedroom-256""" UpperCAmelCase = UNetaDModel.from_pretrained(__lowerCamelCase ) UpperCAmelCase = DDIMScheduler.from_pretrained(__lowerCamelCase ) UpperCAmelCase = DDIMPipeline(unet=__lowerCamelCase , scheduler=__lowerCamelCase ) ddpm.to(__lowerCamelCase ) ddpm.set_progress_bar_config(disable=__lowerCamelCase ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = ddpm(generator=__lowerCamelCase , output_type="""numpy""" ).images UpperCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 2_5_6, 2_5_6, 3) UpperCAmelCase = np.array([0.0_060, 0.0_201, 0.0_344, 0.0_024, 0.0_018, 0.0_002, 0.0_022, 0.0_000, 0.0_069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
1
def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: UpperCAmelCase = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
627
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
1
from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __lowercase : def __init__( self : Any , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any=3 , __lowerCamelCase : int=3_2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Dict=1_0 , __lowerCamelCase : List[Any]=[1_0, 2_0, 3_0, 4_0] , __lowerCamelCase : int=[1, 1, 2, 1] , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : int="relu" , __lowerCamelCase : Dict=3 , __lowerCamelCase : List[str]=None , ) -> int: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = num_channels UpperCAmelCase = embeddings_size UpperCAmelCase = hidden_sizes UpperCAmelCase = depths UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_act UpperCAmelCase = num_labels UpperCAmelCase = scope UpperCAmelCase = len(__lowerCamelCase ) def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Any ) -> Any: """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def _lowercase ( self : Dict , __lowerCamelCase : Dict , __lowerCamelCase : Dict , __lowerCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFRegNetModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , training=__lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def _lowercase ( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Any ) -> List[str]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = TFRegNetForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase , training=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : List[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFRegNetModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def _lowercase ( self : str ) -> List[Any]: """simple docstring""" super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" pass def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" def check_hidden_states_output(__lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] ): UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) , training=__lowerCamelCase ) UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(__lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCAmelCase = layer_type UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> Dict: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(__lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : List[Any] , __lowerCamelCase : List[str]={} ): UpperCAmelCase = model(__lowerCamelCase , return_dict=__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , return_dict=__lowerCamelCase , **__lowerCamelCase ).to_tuple() def recursive_check(__lowerCamelCase : List[Any] , __lowerCamelCase : Dict ): if isinstance(__lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(__lowerCamelCase , __lowerCamelCase ): recursive_check(__lowerCamelCase , __lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(__lowerCamelCase , __lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F""" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}""" ) , ) recursive_check(__lowerCamelCase , __lowerCamelCase ) for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , {"""output_hidden_states""": True} ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) UpperCAmelCase = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) check_equivalence(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , {"""output_hidden_states""": True} ) def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def _lowercase ( self : List[Any] ) -> str: """simple docstring""" for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFRegNetModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Union[str, Any]: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : str ) -> Tuple: """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" UpperCAmelCase = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase , training=__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-0.4_180, -1.5_051, -3.4_836] ) tf.debugging.assert_near(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 )
627
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __a = logging.get_logger(__name__) __a = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """ctc_proj""", """mask_emb""": """masked_spec_embed""", } __a = [ """ctc_proj""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: for attribute in key.split(""".""" ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models UpperCAmelCase = """lm_head""" UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) if weight_type is not None: UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ ).shape else: UpperCAmelCase = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase = value elif weight_type == "weight_g": UpperCAmelCase = value elif weight_type == "weight_v": UpperCAmelCase = value elif weight_type == "bias": UpperCAmelCase = value else: UpperCAmelCase = value logger.info(F"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: UpperCAmelCase = [] UpperCAmelCase = fairseq_model.state_dict() UpperCAmelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase = False if "conv_layers" in name: load_conv_layer( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , hf_model.config.feat_extract_norm == """group""" , ) UpperCAmelCase = True else: for key, mapped_key in MAPPING.items(): UpperCAmelCase = """unispeech.""" + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: UpperCAmelCase = True if "*" in mapped_key: UpperCAmelCase = name.split(lowerCAmelCase_ )[0].split(""".""" )[-2] UpperCAmelCase = mapped_key.replace("""*""" , lowerCAmelCase_ ) if "weight_g" in name: UpperCAmelCase = """weight_g""" elif "weight_v" in name: UpperCAmelCase = """weight_v""" elif "bias" in name: UpperCAmelCase = """bias""" elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase = """weight""" else: UpperCAmelCase = None set_recursively(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) continue if not is_used: unused_weights.append(lowerCAmelCase_ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Optional[int]: UpperCAmelCase = full_name.split("""conv_layers.""" )[-1] UpperCAmelCase = name.split(""".""" ) UpperCAmelCase = int(items[0] ) UpperCAmelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) UpperCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowerCAmelCase_ ) @torch.no_grad() def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=True ) ->List[Any]: if config_path is not None: UpperCAmelCase = UniSpeechConfig.from_pretrained(lowerCAmelCase_ ) else: UpperCAmelCase = UniSpeechConfig() if is_finetuned: if dict_path: UpperCAmelCase = Dictionary.load_from_json(lowerCAmelCase_ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase = target_dict.pad_index UpperCAmelCase = target_dict.bos_index UpperCAmelCase = target_dict.eos_index UpperCAmelCase = len(target_dict.symbols ) UpperCAmelCase = os.path.join(lowerCAmelCase_ , """vocab.json""" ) if not os.path.isdir(lowerCAmelCase_ ): logger.error("""--pytorch_dump_folder_path ({}) should be a directory""".format(lowerCAmelCase_ ) ) return os.makedirs(lowerCAmelCase_ , exist_ok=lowerCAmelCase_ ) UpperCAmelCase = target_dict.indices # fairseq has the <pad> and <s> switched UpperCAmelCase = 4_2 UpperCAmelCase = 4_3 with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" ) as vocab_handle: json.dump(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = WavaVecaPhonemeCTCTokenizer( lowerCAmelCase_ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="""|""" , do_lower_case=lowerCAmelCase_ , ) UpperCAmelCase = True if config.feat_extract_norm == """layer""" else False UpperCAmelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , ) UpperCAmelCase = WavaVecaProcessor(feature_extractor=lowerCAmelCase_ , tokenizer=lowerCAmelCase_ ) processor.save_pretrained(lowerCAmelCase_ ) UpperCAmelCase = UniSpeechForCTC(lowerCAmelCase_ ) else: UpperCAmelCase = UniSpeechForPreTraining(lowerCAmelCase_ ) if is_finetuned: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] ), """w2v_path""": checkpoint_path} ) else: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) UpperCAmelCase = model[0].eval() recursively_load_weights(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) hf_unispeech.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) __a = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
627
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
1
from __future__ import annotations import requests def _UpperCamelCase ( lowerCAmelCase_ ) ->dict: UpperCAmelCase = F"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowerCAmelCase_ ).json() def _UpperCamelCase ( lowerCAmelCase_ = 1_0 ) ->list[dict]: UpperCAmelCase = """https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty""" UpperCAmelCase = requests.get(lowerCAmelCase_ ).json()[:max_stories] return [get_hackernews_story(lowerCAmelCase_ ) for story_id in story_ids] def _UpperCamelCase ( lowerCAmelCase_ = 1_0 ) ->str: UpperCAmelCase = hackernews_top_stories(lowerCAmelCase_ ) return "\n".join("""* [{title}]({url})""".format(**lowerCAmelCase_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
627
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
1
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer __a = logging.get_logger(__name__) class __lowercase ( __snake_case ): UpperCamelCase = '''AutoTokenizer''' UpperCamelCase = ['''tokenizer'''] UpperCamelCase = { '''semantic_prompt''': 1, '''coarse_prompt''': 2, '''fine_prompt''': 2, } def __init__( self : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any=None ) -> Any: """simple docstring""" super().__init__(__lowerCamelCase ) UpperCAmelCase = speaker_embeddings @classmethod def _lowercase ( cls : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : str="speaker_embeddings_path.json" , **__lowerCamelCase : Any ) -> str: """simple docstring""" if speaker_embeddings_dict_path is not None: UpperCAmelCase = get_file_from_repo( __lowerCamelCase , __lowerCamelCase , subfolder=kwargs.pop("""subfolder""" , __lowerCamelCase ) , cache_dir=kwargs.pop("""cache_dir""" , __lowerCamelCase ) , force_download=kwargs.pop("""force_download""" , __lowerCamelCase ) , proxies=kwargs.pop("""proxies""" , __lowerCamelCase ) , resume_download=kwargs.pop("""resume_download""" , __lowerCamelCase ) , local_files_only=kwargs.pop("""local_files_only""" , __lowerCamelCase ) , use_auth_token=kwargs.pop("""use_auth_token""" , __lowerCamelCase ) , revision=kwargs.pop("""revision""" , __lowerCamelCase ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(__lowerCamelCase , __lowerCamelCase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) UpperCAmelCase = None else: with open(__lowerCamelCase ) as speaker_embeddings_json: UpperCAmelCase = json.load(__lowerCamelCase ) else: UpperCAmelCase = None UpperCAmelCase = AutoTokenizer.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) return cls(tokenizer=__lowerCamelCase , speaker_embeddings=__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int="speaker_embeddings_path.json" , __lowerCamelCase : Any="speaker_embeddings" , __lowerCamelCase : bool = False , **__lowerCamelCase : Any , ) -> str: """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(__lowerCamelCase , __lowerCamelCase , """v2""" ) , exist_ok=__lowerCamelCase ) UpperCAmelCase = {} UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": UpperCAmelCase = self._load_voice_preset(__lowerCamelCase ) UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["""repo_or_path"""] , __lowerCamelCase , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=__lowerCamelCase , ) UpperCAmelCase = os.path.join(__lowerCamelCase , F"""{prompt_key}_{key}.npy""" ) UpperCAmelCase = tmp_dict with open(os.path.join(__lowerCamelCase , __lowerCamelCase ) , """w""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) super().save_pretrained(__lowerCamelCase , __lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : int , __lowerCamelCase : str = None , **__lowerCamelCase : Tuple ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.speaker_embeddings[voice_preset] UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get("""repo_or_path""" , """/""" ) , voice_preset_paths[key] , subfolder=kwargs.pop("""subfolder""" , __lowerCamelCase ) , cache_dir=kwargs.pop("""cache_dir""" , __lowerCamelCase ) , force_download=kwargs.pop("""force_download""" , __lowerCamelCase ) , proxies=kwargs.pop("""proxies""" , __lowerCamelCase ) , resume_download=kwargs.pop("""resume_download""" , __lowerCamelCase ) , local_files_only=kwargs.pop("""local_files_only""" , __lowerCamelCase ) , use_auth_token=kwargs.pop("""use_auth_token""" , __lowerCamelCase ) , revision=kwargs.pop("""revision""" , __lowerCamelCase ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) UpperCAmelCase = np.load(__lowerCamelCase ) return voice_preset_dict def _lowercase ( self : Any , __lowerCamelCase : Optional[dict] = None ) -> Any: """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self : int , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[str]=None , __lowerCamelCase : str="pt" , __lowerCamelCase : Any=2_5_6 , __lowerCamelCase : List[str]=False , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[Any]=False , **__lowerCamelCase : Optional[int] , ) -> List[Any]: """simple docstring""" if voice_preset is not None and not isinstance(__lowerCamelCase , __lowerCamelCase ): if ( isinstance(__lowerCamelCase , __lowerCamelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): UpperCAmelCase = self._load_voice_preset(__lowerCamelCase ) else: if isinstance(__lowerCamelCase , __lowerCamelCase ) and not voice_preset.endswith(""".npz""" ): UpperCAmelCase = voice_preset + """.npz""" UpperCAmelCase = np.load(__lowerCamelCase ) if voice_preset is not None: self._validate_voice_preset_dict(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = BatchFeature(data=__lowerCamelCase , tensor_type=__lowerCamelCase ) UpperCAmelCase = self.tokenizer( __lowerCamelCase , return_tensors=__lowerCamelCase , padding="""max_length""" , max_length=__lowerCamelCase , return_attention_mask=__lowerCamelCase , return_token_type_ids=__lowerCamelCase , add_special_tokens=__lowerCamelCase , **__lowerCamelCase , ) if voice_preset is not None: UpperCAmelCase = voice_preset return encoded_text
627
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
1
from queue import PriorityQueue from typing import Any import numpy as np def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->float | int: for nxt, d in graph[v]: if nxt in visited_forward: continue UpperCAmelCase = cst_fwd.get(lowerCAmelCase_ , np.inf ) UpperCAmelCase = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) UpperCAmelCase = new_cost_f UpperCAmelCase = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: UpperCAmelCase = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->int: UpperCAmelCase = -1 UpperCAmelCase = set() UpperCAmelCase = set() UpperCAmelCase = {source: 0} UpperCAmelCase = {destination: 0} UpperCAmelCase = {source: None} UpperCAmelCase = {destination: None} UpperCAmelCase = PriorityQueue() UpperCAmelCase = PriorityQueue() UpperCAmelCase = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): UpperCAmelCase , UpperCAmelCase = queue_forward.get() visited_forward.add(lowerCAmelCase_ ) UpperCAmelCase , UpperCAmelCase = queue_backward.get() visited_backward.add(lowerCAmelCase_ ) UpperCAmelCase = pass_and_relaxation( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) UpperCAmelCase = pass_and_relaxation( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: UpperCAmelCase = shortest_distance return shortest_path_distance __a = { """B""": [["""C""", 1]], """C""": [["""D""", 1]], """D""": [["""F""", 1]], """E""": [["""B""", 1], ["""G""", 2]], """F""": [], """G""": [["""F""", 1]], } __a = { """B""": [["""E""", 1]], """C""": [["""B""", 1]], """D""": [["""C""", 1]], """F""": [["""D""", 1], ["""G""", 1]], """E""": [[None, np.inf]], """G""": [["""E""", 2]], } if __name__ == "__main__": import doctest doctest.testmod()
627
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
1
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = LongformerTokenizer UpperCamelCase = True UpperCamelCase = LongformerTokenizerFast UpperCamelCase = True def _lowercase ( self : Optional[Any] ) -> str: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) def _lowercase ( self : Dict , **__lowerCamelCase : Optional[Any] ) -> Dict: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Tuple , **__lowerCamelCase : Any ) -> Dict: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Tuple , __lowerCamelCase : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = """lower newer""" UpperCAmelCase = """lower newer""" return input_text, output_text def _lowercase ( self : Tuple ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCAmelCase = """lower newer""" UpperCAmelCase = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] UpperCAmelCase = tokenizer.tokenize(__lowerCamelCase ) # , add_prefix_space=True) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = tokens + [tokenizer.unk_token] UpperCAmelCase = [0, 1, 2, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() self.assertListEqual(tokenizer.encode("""Hello world!""" , add_special_tokens=__lowerCamelCase ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 2] ) self.assertListEqual( tokenizer.encode("""Hello world! cécé herlolip 418""" , add_special_tokens=__lowerCamelCase ) , [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2] , ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.tokenizer_class.from_pretrained("""allenai/longformer-base-4096""" ) UpperCAmelCase = tokenizer.encode("""sequence builders""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode("""multi-sequence build""" , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode( """sequence builders""" , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.encode( """sequence builders""" , """multi-sequence build""" , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ) UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase , __lowerCamelCase ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = """Encode this sequence.""" UpperCAmelCase = tokenizer.byte_encoder[""" """.encode("""utf-8""" )[0]] # Testing encoder arguments UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase , add_prefix_space=__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(__lowerCamelCase , __lowerCamelCase ) tokenizer.add_special_tokens({"""bos_token""": """<s>"""} ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(__lowerCamelCase , __lowerCamelCase ) # Testing spaces after special tokens UpperCAmelCase = """<mask>""" tokenizer.add_special_tokens( {"""mask_token""": AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase )} ) # mask token has a left space UpperCAmelCase = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) UpperCAmelCase = """Encode <mask> sequence""" UpperCAmelCase = """Encode <mask>sequence""" UpperCAmelCase = tokenizer.encode(__lowerCamelCase ) UpperCAmelCase = encoded.index(__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = tokenizer.encode(__lowerCamelCase ) UpperCAmelCase = encoded.index(__lowerCamelCase ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" pass def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = self.tokenizer_class.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) UpperCAmelCase = """A, <mask> AllenNLP sentence.""" UpperCAmelCase = tokenizer_r.encode_plus(__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_token_type_ids=__lowerCamelCase ) UpperCAmelCase = tokenizer_p.encode_plus(__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_token_type_ids=__lowerCamelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ) , sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ) , sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ) , ) UpperCAmelCase = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) UpperCAmelCase = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""] , [0, 2_5_0, 6, 5_0_2_6_4, 3_8_2_3, 4_8_7, 2_1_9_9_2, 3_6_4_5, 4, 2] ) self.assertSequenceEqual( __lowerCamelCase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( __lowerCamelCase , ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) UpperCAmelCase = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["""add_prefix_space"""] , __lowerCamelCase ) self.assertEqual(post_processor_state["""add_prefix_space"""] , __lowerCamelCase ) self.assertEqual(post_processor_state["""trim_offsets"""] , __lowerCamelCase ) def _lowercase ( self : Any ) -> Tuple: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCAmelCase = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` UpperCAmelCase = F"""{text_of_1_token} {text_of_1_token}""" UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ) + 1, len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ) + 1, len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ), len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__lowerCamelCase ), len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = F""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__lowerCamelCase ) + 1, 1 + len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__lowerCamelCase ), 1 + len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , ) UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( __lowerCamelCase , use_fast=__lowerCamelCase , add_prefix_space=__lowerCamelCase , trim_offsets=__lowerCamelCase ) UpperCAmelCase = tokenizer_r(__lowerCamelCase , return_offsets_mapping=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(__lowerCamelCase )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__lowerCamelCase ), 1 + len(__lowerCamelCase ) + 1 + len(__lowerCamelCase )) , )
627
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
1
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __a = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = XLMProphetNetTokenizer UpperCamelCase = False UpperCamelCase = True def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XLMProphetNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = """[PAD]""" UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """[PAD]""" ) self.assertEqual(vocab_keys[1] , """[CLS]""" ) self.assertEqual(vocab_keys[-1] , """j""" ) self.assertEqual(len(__lowerCamelCase ) , 1_0_1_2 ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_2 ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = XLMProphetNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) UpperCAmelCase = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__lowerCamelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) UpperCAmelCase = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, -9, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, -9, 4] ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """[UNK]""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """[UNK]""", """.""", ] , ) @cached_property def _lowercase ( self : int ) -> Any: """simple docstring""" return XLMProphetNetTokenizer.from_pretrained("""microsoft/xprophetnet-large-wiki100-cased""" ) @slow def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = """Hello World!""" UpperCAmelCase = [3_5_3_8_9, 6_6_7_2, 4_9, 2] self.assertListEqual(__lowerCamelCase , self.big_tokenizer.encode(__lowerCamelCase ) ) @slow def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = {"""input_ids""": [[1_1_0_7_3, 8_2_7_8_3, 1_8, 2_6, 8_2_7_8_3, 5_4_9, 5_1_5_4_0, 2_4_8, 1_7_2_0_9, 1_3_0_1, 2_1_7, 2_0, 2_1_5_1_8_6, 1_3_2_5, 1_4_7, 1_7_2_0_9, 1_3_0_1, 2_1_7, 2_0, 5_6_3_7_0, 5_3, 1_2_2_0_2_0, 2_0, 1_6_4_7_7, 2_7, 8_7_3_5_5, 4_5_4_8, 2_0, 4_7_2_8, 7_8_3_9_2, 1_7, 1_5_9_9_6_9, 1_8, 2_6, 2_4_4_9_1, 6_2_9, 1_5, 5_3_8, 2_2_7_0_4, 5_4_3_9, 1_5, 2_7_8_8, 2_4_4_9_1, 9_8_8_5, 1_5, 4_3_5_3_4, 6_0_5, 1_5, 8_1_4, 1_8_4_0_3, 3_3_2_0_0, 2_9, 1_5, 4_3_5_3_4, 2_4_4_5_8, 1_2_4_1_0, 1_1_1, 2_4_9_6_6, 8_3_6_6_9, 9_6_3_7, 1_4_4_0_6_8, 2_6, 8_5_0, 2_2_3_4_6, 2_7, 1_4_7, 2_4_9_6_6, 8_3_6_6_9, 8_3_4_9_0, 2_6, 3_9_1_1_3, 7_3_5, 2_7, 6_8_9, 6_5_6, 2_8_0_0, 1_3_3_9, 4_6_0_0, 5_3, 1_2_2_0_2_0, 1_1_5_7_8_5, 3_4, 8_1_6, 1_3_3_9, 4_6_8_8_7, 1_8, 1_4_7, 5_3_9_0_5, 1_9_5_1, 4_2_2_3_8, 4_1_1_7_0, 1_7_7_3_2, 8_3_4, 4_3_6, 1_5, 2_7_5_2_3, 9_8_7_3_3, 2_1_7, 1_4_7, 5_5_4_2, 4_9_8_1, 9_3_0, 1_7_3_4_7, 1_6, 2], [2_0_0_9_1, 6_2_9, 9_4, 8_2_7_8_6, 5_8, 4_9_0, 2_0, 1_5_2_8, 8_4, 5_3_9_0_5, 3_4_4, 8_0_5_9_2, 1_1_0_1_2_8, 1_8_8_2_2, 5_2_6_7, 1_3_0_6, 6_2, 1_5_2_5_3_7, 3_0_8, 7_9_9_7, 4_0_1, 1_2_4_4_2_7, 5_4_9, 3_5_4_4_2, 2_2_5, 1_0_9, 1_5_0_5_5, 2_5_7_4_8, 1_4_7, 7_1_1_9, 4_3_7_1_2, 3_4, 7_6_7, 1_3_5_3_6_6, 1_8, 1_6, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_9_2, 6_3_7_8_4, 1_1_9_4_6_6, 1_7, 1_4_7_8_0_8, 8_8_2_1_4, 1_8, 6_5_6, 8_1, 3_2, 3_2_9_6, 1_0_2_8_0, 1_6, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name="""microsoft/xprophetnet-large-wiki100-cased""" , revision="""1acad1643ddd54a44df6a1b797ada8373685d90e""" , )
627
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
1
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __a = Mapping[str, np.ndarray] __a = Mapping[str, Any] # Is a nested dict. __a = 0.01 @dataclasses.dataclass(frozen=__snake_case ) class __lowercase : UpperCamelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. UpperCamelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. UpperCamelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. UpperCamelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. UpperCamelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions UpperCamelCase = None # Optional remark about the protein. Included as a comment in output PDB # files UpperCamelCase = None # Templates used to generate this protein (prediction-only) UpperCamelCase = None # Chain corresponding to each parent UpperCamelCase = None def _UpperCamelCase ( lowerCAmelCase_ ) ->Protein: UpperCAmelCase = R"""(\[[A-Z]+\]\n)""" UpperCAmelCase = [tag.strip() for tag in re.split(lowerCAmelCase_ , lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0] UpperCAmelCase = zip(tags[0::2] , [l.split("""\n""" ) for l in tags[1::2]] ) UpperCAmelCase = ["N", "CA", "C"] UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None for g in groups: if "[PRIMARY]" == g[0]: UpperCAmelCase = g[1][0].strip() for i in range(len(lowerCAmelCase_ ) ): if seq[i] not in residue_constants.restypes: UpperCAmelCase = """X""" # FIXME: strings are immutable UpperCAmelCase = np.array( [residue_constants.restype_order.get(lowerCAmelCase_ , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: UpperCAmelCase = [] for axis in range(3 ): tertiary.append(list(map(lowerCAmelCase_ , g[1][axis].split() ) ) ) UpperCAmelCase = np.array(lowerCAmelCase_ ) UpperCAmelCase = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(lowerCAmelCase_ ): UpperCAmelCase = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: UpperCAmelCase = np.array(list(map({"""-""": 0, """+""": 1}.get , g[1][0].strip() ) ) ) UpperCAmelCase = np.zeros( ( len(lowerCAmelCase_ ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(lowerCAmelCase_ ): UpperCAmelCase = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=lowerCAmelCase_ , atom_mask=lowerCAmelCase_ , aatype=lowerCAmelCase_ , residue_index=np.arange(len(lowerCAmelCase_ ) ) , b_factors=lowerCAmelCase_ , ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = 0 ) ->List[str]: UpperCAmelCase = [] UpperCAmelCase = prot.remark if remark is not None: pdb_headers.append(F"""REMARK {remark}""" ) UpperCAmelCase = prot.parents UpperCAmelCase = prot.parents_chain_index if parents is not None and parents_chain_index is not None: UpperCAmelCase = [p for i, p in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if i == chain_id] if parents is None or len(lowerCAmelCase_ ) == 0: UpperCAmelCase = ["""N/A"""] pdb_headers.append(F"""PARENT {' '.join(lowerCAmelCase_ )}""" ) return pdb_headers def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str: UpperCAmelCase = [] UpperCAmelCase = pdb_str.split("""\n""" ) UpperCAmelCase = prot.remark if remark is not None: out_pdb_lines.append(F"""REMARK {remark}""" ) UpperCAmelCase = 42 if prot.parents is not None and len(prot.parents ) > 0: UpperCAmelCase = [] if prot.parents_chain_index is not None: UpperCAmelCase = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(lowerCAmelCase_ ) , [] ) parent_dict[str(lowerCAmelCase_ )].append(lowerCAmelCase_ ) UpperCAmelCase = max([int(lowerCAmelCase_ ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): UpperCAmelCase = parent_dict.get(str(lowerCAmelCase_ ) , ["""N/A"""] ) parents_per_chain.append(lowerCAmelCase_ ) else: parents_per_chain.append(list(prot.parents ) ) else: UpperCAmelCase = [["""N/A"""]] def make_parent_line(lowerCAmelCase_ ) -> str: return F"""PARENT {' '.join(lowerCAmelCase_ )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) UpperCAmelCase = 0 for i, l in enumerate(lowerCAmelCase_ ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(lowerCAmelCase_ ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(lowerCAmelCase_ ): UpperCAmelCase = parents_per_chain[chain_counter] else: UpperCAmelCase = ["""N/A"""] out_pdb_lines.append(make_parent_line(lowerCAmelCase_ ) ) return "\n".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = residue_constants.restypes + ["""X"""] def res_atoa(lowerCAmelCase_ ) -> str: return residue_constants.restype_atoa.get(restypes[r] , """UNK""" ) UpperCAmelCase = residue_constants.atom_types UpperCAmelCase = [] UpperCAmelCase = prot.atom_mask UpperCAmelCase = prot.aatype UpperCAmelCase = prot.atom_positions UpperCAmelCase = prot.residue_index.astype(np.intaa ) UpperCAmelCase = prot.b_factors UpperCAmelCase = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("""Invalid aatypes.""" ) UpperCAmelCase = get_pdb_headers(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: pdb_lines.extend(lowerCAmelCase_ ) UpperCAmelCase = aatype.shape[0] UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = string.ascii_uppercase UpperCAmelCase = None # Add all atom sites. for i in range(lowerCAmelCase_ ): UpperCAmelCase = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(lowerCAmelCase_ , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue UpperCAmelCase = """ATOM""" UpperCAmelCase = atom_name if len(lowerCAmelCase_ ) == 4 else F""" {atom_name}""" UpperCAmelCase = """""" UpperCAmelCase = """""" UpperCAmelCase = 1.00 UpperCAmelCase = atom_name[0] # Protein supports only C, N, O, S, this works. UpperCAmelCase = """""" UpperCAmelCase = """A""" if chain_index is not None: UpperCAmelCase = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! UpperCAmelCase = ( F"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" F"""{res_name_a:>3} {chain_tag:>1}""" F"""{residue_index[i]:>4}{insertion_code:>1} """ F"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" F"""{occupancy:>6.2f}{b_factor:>6.2f} """ F"""{element:>2}{charge:>2}""" ) pdb_lines.append(lowerCAmelCase_ ) atom_index += 1 UpperCAmelCase = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: UpperCAmelCase = True UpperCAmelCase = chain_index[i + 1] if should_terminate: # Close the chain. UpperCAmelCase = """TER""" UpperCAmelCase = ( F"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(lowerCAmelCase_ ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(lowerCAmelCase_ , lowerCAmelCase_ ) ) pdb_lines.append("""END""" ) pdb_lines.append("""""" ) return "\n".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->np.ndarray: return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , ) ->Protein: return Protein( aatype=features["""aatype"""] , atom_positions=result["""final_atom_positions"""] , atom_mask=result["""final_atom_mask"""] , residue_index=features["""residue_index"""] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["""final_atom_mask"""] ) , chain_index=lowerCAmelCase_ , remark=lowerCAmelCase_ , parents=lowerCAmelCase_ , parents_chain_index=lowerCAmelCase_ , )
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/timesformer""": """https://huggingface.co/facebook/timesformer/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''timesformer''' def __init__( self : Union[str, Any] , __lowerCamelCase : Any=2_2_4 , __lowerCamelCase : Optional[Any]=1_6 , __lowerCamelCase : int=3 , __lowerCamelCase : Dict=8 , __lowerCamelCase : Dict=7_6_8 , __lowerCamelCase : int=1_2 , __lowerCamelCase : Any=1_2 , __lowerCamelCase : Dict=3_0_7_2 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : Any=0.0 , __lowerCamelCase : Dict=0.0 , __lowerCamelCase : Dict=0.02 , __lowerCamelCase : int=1e-6 , __lowerCamelCase : Tuple=True , __lowerCamelCase : List[Any]="divided_space_time" , __lowerCamelCase : str=0 , **__lowerCamelCase : str , ) -> str: """simple docstring""" super().__init__(**__lowerCamelCase ) UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = num_frames UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = qkv_bias UpperCAmelCase = attention_type UpperCAmelCase = drop_path_rate
627
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
1
from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP __a = logging.get_logger(__name__) # pylint: disable=invalid-name __a = """ Examples: ```py >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline >>> import torch >>> pipe_prior = KandinskyPriorPipeline.from_pretrained(\"kandinsky-community/Kandinsky-2-1-prior\") >>> pipe_prior.to(\"cuda\") >>> prompt = \"red cat, 4k photo\" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> negative_image_emb = out.negative_image_embeds >>> pipe = KandinskyPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-1\") >>> pipe.to(\"cuda\") >>> image = pipe( ... prompt, ... image_embeds=image_emb, ... negative_image_embeds=negative_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... ).images >>> image[0].save(\"cat.png\") ``` """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=8 ) ->List[str]: UpperCAmelCase = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 UpperCAmelCase = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class __lowercase ( __snake_case ): def __init__( self : Optional[Any] , __lowerCamelCase : MultilingualCLIP , __lowerCamelCase : XLMRobertaTokenizer , __lowerCamelCase : UNetaDConditionModel , __lowerCamelCase : Union[DDIMScheduler, DDPMScheduler] , __lowerCamelCase : VQModel , ) -> List[str]: """simple docstring""" super().__init__() self.register_modules( text_encoder=__lowerCamelCase , tokenizer=__lowerCamelCase , unet=__lowerCamelCase , scheduler=__lowerCamelCase , movq=__lowerCamelCase , ) UpperCAmelCase = 2 ** (len(self.movq.config.block_out_channels ) - 1) def _lowercase ( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Any , __lowerCamelCase : str , __lowerCamelCase : Any ) -> Tuple: """simple docstring""" if latents is None: UpperCAmelCase = randn_tensor(__lowerCamelCase , generator=__lowerCamelCase , device=__lowerCamelCase , dtype=__lowerCamelCase ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase = latents.to(__lowerCamelCase ) UpperCAmelCase = latents * scheduler.init_noise_sigma return latents def _lowercase ( self : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : Any , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any]=None , ) -> List[str]: """simple docstring""" UpperCAmelCase = len(__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else 1 # get prompt text embeddings UpperCAmelCase = self.tokenizer( __lowerCamelCase , padding="""max_length""" , truncation=__lowerCamelCase , max_length=7_7 , return_attention_mask=__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_tensors="""pt""" , ) UpperCAmelCase = text_inputs.input_ids UpperCAmelCase = self.tokenizer(__lowerCamelCase , padding="""longest""" , return_tensors="""pt""" ).input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" F""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) UpperCAmelCase = text_input_ids.to(__lowerCamelCase ) UpperCAmelCase = text_inputs.attention_mask.to(__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = self.text_encoder( input_ids=__lowerCamelCase , attention_mask=__lowerCamelCase ) UpperCAmelCase = prompt_embeds.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = text_encoder_hidden_states.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = text_mask.repeat_interleave(__lowerCamelCase , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase = 42 if negative_prompt is None: UpperCAmelCase = [""""""] * batch_size elif type(__lowerCamelCase ) is not type(__lowerCamelCase ): raise TypeError( F"""`negative_prompt` should be the same type to `prompt`, but got {type(__lowerCamelCase )} !=""" F""" {type(__lowerCamelCase )}.""" ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = [negative_prompt] elif batch_size != len(__lowerCamelCase ): raise ValueError( F"""`negative_prompt`: {negative_prompt} has batch size {len(__lowerCamelCase )}, but `prompt`:""" F""" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches""" """ the batch size of `prompt`.""" ) else: UpperCAmelCase = negative_prompt UpperCAmelCase = self.tokenizer( __lowerCamelCase , padding="""max_length""" , max_length=7_7 , truncation=__lowerCamelCase , return_attention_mask=__lowerCamelCase , add_special_tokens=__lowerCamelCase , return_tensors="""pt""" , ) UpperCAmelCase = uncond_input.input_ids.to(__lowerCamelCase ) UpperCAmelCase = uncond_input.attention_mask.to(__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = self.text_encoder( input_ids=__lowerCamelCase , attention_mask=__lowerCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method UpperCAmelCase = negative_prompt_embeds.shape[1] UpperCAmelCase = negative_prompt_embeds.repeat(1 , __lowerCamelCase ) UpperCAmelCase = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __lowerCamelCase ) UpperCAmelCase = uncond_text_encoder_hidden_states.shape[1] UpperCAmelCase = uncond_text_encoder_hidden_states.repeat(1 , __lowerCamelCase , 1 ) UpperCAmelCase = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt , __lowerCamelCase , -1 ) UpperCAmelCase = uncond_text_mask.repeat_interleave(__lowerCamelCase , dim=0 ) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase = torch.cat([negative_prompt_embeds, prompt_embeds] ) UpperCAmelCase = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states] ) UpperCAmelCase = torch.cat([uncond_text_mask, text_mask] ) return prompt_embeds, text_encoder_hidden_states, text_mask def _lowercase ( self : Tuple , __lowerCamelCase : Dict=0 ) -> Any: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) UpperCAmelCase = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any]=0 ) -> List[str]: """simple docstring""" if is_accelerate_available() and is_accelerate_version(""">=""" , """0.17.0.dev0""" ): from accelerate import cpu_offload_with_hook else: raise ImportError("""`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.""" ) UpperCAmelCase = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("""cpu""" , silence_dtype_warnings=__lowerCamelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: UpperCAmelCase , UpperCAmelCase = cpu_offload_with_hook(__lowerCamelCase , __lowerCamelCase , prev_module_hook=__lowerCamelCase ) if self.safety_checker is not None: UpperCAmelCase , UpperCAmelCase = cpu_offload_with_hook(self.safety_checker , __lowerCamelCase , prev_module_hook=__lowerCamelCase ) # We'll offload the last model manually. UpperCAmelCase = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" if not hasattr(self.unet , """_hf_hook""" ): return self.device for module in self.unet.modules(): if ( hasattr(__lowerCamelCase , """_hf_hook""" ) and hasattr(module._hf_hook , """execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(__lowerCamelCase ) def __call__( self : List[str] , __lowerCamelCase : Union[str, List[str]] , __lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , __lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , __lowerCamelCase : Optional[Union[str, List[str]]] = None , __lowerCamelCase : int = 5_1_2 , __lowerCamelCase : int = 5_1_2 , __lowerCamelCase : int = 1_0_0 , __lowerCamelCase : float = 4.0 , __lowerCamelCase : int = 1 , __lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __lowerCamelCase : Optional[torch.FloatTensor] = None , __lowerCamelCase : Optional[str] = "pil" , __lowerCamelCase : bool = True , ) -> Any: """simple docstring""" if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = 1 elif isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = len(__lowerCamelCase ) else: raise ValueError(F"""`prompt` has to be of type `str` or `list` but is {type(__lowerCamelCase )}""" ) UpperCAmelCase = self._execution_device UpperCAmelCase = batch_size * num_images_per_prompt UpperCAmelCase = guidance_scale > 1.0 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self._encode_prompt( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = torch.cat(__lowerCamelCase , dim=0 ) if isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = torch.cat(__lowerCamelCase , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase = image_embeds.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = negative_image_embeds.repeat_interleave(__lowerCamelCase , dim=0 ) UpperCAmelCase = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to( dtype=prompt_embeds.dtype , device=__lowerCamelCase ) self.scheduler.set_timesteps(__lowerCamelCase , device=__lowerCamelCase ) UpperCAmelCase = self.scheduler.timesteps UpperCAmelCase = self.unet.config.in_channels UpperCAmelCase , UpperCAmelCase = get_new_h_w(__lowerCamelCase , __lowerCamelCase , self.movq_scale_factor ) # create initial latent UpperCAmelCase = self.prepare_latents( (batch_size, num_channels_latents, height, width) , text_encoder_hidden_states.dtype , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , self.scheduler , ) for i, t in enumerate(self.progress_bar(__lowerCamelCase ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase = {"""text_embeds""": prompt_embeds, """image_embeds""": image_embeds} UpperCAmelCase = self.unet( sample=__lowerCamelCase , timestep=__lowerCamelCase , encoder_hidden_states=__lowerCamelCase , added_cond_kwargs=__lowerCamelCase , return_dict=__lowerCamelCase , )[0] if do_classifier_free_guidance: UpperCAmelCase , UpperCAmelCase = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase , UpperCAmelCase = noise_pred.chunk(2 ) UpperCAmelCase , UpperCAmelCase = variance_pred.chunk(2 ) UpperCAmelCase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , """variance_type""" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase , UpperCAmelCase = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase = self.scheduler.step( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase , ).prev_sample # post-processing UpperCAmelCase = self.movq.decode(__lowerCamelCase , force_not_quantize=__lowerCamelCase )["""sample"""] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase = image * 0.5 + 0.5 UpperCAmelCase = image.clamp(0 , 1 ) UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase = self.numpy_to_pil(__lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__lowerCamelCase )
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
import numpy as np class __lowercase : def __init__( self : Union[str, Any] ) -> Dict: """simple docstring""" UpperCAmelCase = (0, 0) UpperCAmelCase = None UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 0 def __eq__( self : Any , __lowerCamelCase : str ) -> Optional[int]: """simple docstring""" return self.position == cell.position def _lowercase ( self : List[Any] ) -> List[str]: """simple docstring""" print(self.position ) class __lowercase : def __init__( self : Any , __lowerCamelCase : Union[str, Any]=(5, 5) ) -> str: """simple docstring""" UpperCAmelCase = np.zeros(__lowerCamelCase ) UpperCAmelCase = world_size[0] UpperCAmelCase = world_size[1] def _lowercase ( self : Any ) -> str: """simple docstring""" print(self.w ) def _lowercase ( self : Dict , __lowerCamelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] UpperCAmelCase = cell.position[0] UpperCAmelCase = cell.position[1] UpperCAmelCase = [] for n in neughbour_cord: UpperCAmelCase = current_x + n[0] UpperCAmelCase = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: UpperCAmelCase = Cell() UpperCAmelCase = (x, y) UpperCAmelCase = cell neighbours.append(__lowerCamelCase ) return neighbours def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Optional[int]: UpperCAmelCase = [] UpperCAmelCase = [] _open.append(lowerCAmelCase_ ) while _open: UpperCAmelCase = np.argmin([n.f for n in _open] ) UpperCAmelCase = _open[min_f] _closed.append(_open.pop(lowerCAmelCase_ ) ) if current == goal: break for n in world.get_neigbours(lowerCAmelCase_ ): for c in _closed: if c == n: continue UpperCAmelCase = current.g + 1 UpperCAmelCase , UpperCAmelCase = n.position UpperCAmelCase , UpperCAmelCase = goal.position UpperCAmelCase = (ya - ya) ** 2 + (xa - xa) ** 2 UpperCAmelCase = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(lowerCAmelCase_ ) UpperCAmelCase = [] while current.parent is not None: path.append(current.position ) UpperCAmelCase = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": __a = Gridworld() # Start position and goal __a = Cell() __a = (0, 0) __a = Cell() __a = (4, 4) print(F"""path from {start.position} to {goal.position}""") __a = astar(world, start, goal) # Just for visual reasons. for i in s: __a = 1 print(world.w)
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow __a = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ """text-classification""", """language-modeling""", """summarization""", """token-classification""", """question-answering""", ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) __a = logging.getLogger() def _UpperCamelCase ( ) ->Union[str, Any]: UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""-f""" ) UpperCAmelCase = parser.parse_args() return args.f def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_="eval" ) ->Optional[int]: UpperCAmelCase = os.path.join(lowerCAmelCase_ , F"""{split}_results.json""" ) if os.path.exists(lowerCAmelCase_ ): with open(lowerCAmelCase_ , """r""" ) as f: return json.load(lowerCAmelCase_ ) raise ValueError(F"""can't find {path}""" ) __a = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __lowercase ( __snake_case ): def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_flax_glue.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_accuracy"""] , 0.75 ) @slow def _lowercase ( self : Optional[int] ) -> int: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_clm_flax.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertLess(result["""eval_perplexity"""] , 1_0_0 ) @slow def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_summarization_flax.main() UpperCAmelCase = get_results(__lowerCamelCase , split="""test""" ) self.assertGreaterEqual(result["""test_rouge1"""] , 1_0 ) self.assertGreaterEqual(result["""test_rouge2"""] , 2 ) self.assertGreaterEqual(result["""test_rougeL"""] , 7 ) self.assertGreaterEqual(result["""test_rougeLsum"""] , 7 ) @slow def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_mlm_flax.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertLess(result["""eval_perplexity"""] , 4_2 ) @slow def _lowercase ( self : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_ta_mlm_flax.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_accuracy"""] , 0.42 ) @slow def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase = 7 if get_gpu_count() > 1 else 2 UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_flax_ner.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_accuracy"""] , 0.75 ) self.assertGreaterEqual(result["""eval_f1"""] , 0.3 ) @slow def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): run_qa.main() UpperCAmelCase = get_results(__lowerCamelCase ) self.assertGreaterEqual(result["""eval_f1"""] , 3_0 ) self.assertGreaterEqual(result["""eval_exact"""] , 3_0 )
627
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
1
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __a = TypeVar("""T""") class __lowercase ( Generic[T] ): def __init__( self : Optional[int] , __lowerCamelCase : T ) -> str: """simple docstring""" UpperCAmelCase = data UpperCAmelCase = None def __str__( self : Dict ) -> str: """simple docstring""" return F"""{self.data}""" class __lowercase ( Generic[T] ): def __init__( self : List[Any] ) -> None: """simple docstring""" UpperCAmelCase = None def __iter__( self : Tuple ) -> Iterator[T]: """simple docstring""" UpperCAmelCase = self.top while node: yield node.data UpperCAmelCase = node.next def __str__( self : Union[str, Any] ) -> str: """simple docstring""" return "->".join([str(__lowerCamelCase ) for item in self] ) def __len__( self : Dict ) -> int: """simple docstring""" return len(tuple(iter(self ) ) ) def _lowercase ( self : Union[str, Any] ) -> bool: """simple docstring""" return self.top is None def _lowercase ( self : Any , __lowerCamelCase : T ) -> None: """simple docstring""" UpperCAmelCase = Node(__lowerCamelCase ) if not self.is_empty(): UpperCAmelCase = self.top UpperCAmelCase = node def _lowercase ( self : Dict ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""pop from empty stack""" ) assert isinstance(self.top , __lowerCamelCase ) UpperCAmelCase = self.top UpperCAmelCase = self.top.next return pop_node.data def _lowercase ( self : str ) -> T: """simple docstring""" if self.is_empty(): raise IndexError("""peek from empty stack""" ) assert self.top is not None return self.top.data def _lowercase ( self : int ) -> None: """simple docstring""" UpperCAmelCase = None if __name__ == "__main__": from doctest import testmod testmod()
627
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
1
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse("""3.8"""): import importlib_metadata else: import importlib.metadata as importlib_metadata __a = """""" if version.parse(importlib_metadata.version("""jiwer""")) < version.parse("""2.3.0"""): class __lowercase ( tr.AbstractTransform ): def __init__( self : Any , __lowerCamelCase : str = " " ) -> List[Any]: """simple docstring""" UpperCAmelCase = sentence_delimiter def _lowercase ( self : List[Any] , __lowerCamelCase : str ) -> int: """simple docstring""" return list(__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : List[str] ) -> List[str]: """simple docstring""" UpperCAmelCase = [] for sent_idx, sentence in enumerate(__lowerCamelCase ): chars.extend(self.process_string(__lowerCamelCase ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(__lowerCamelCase ) - 1: chars.append(self.sentence_delimiter ) return chars __a = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: __a = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) __a = """\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ __a = """\ Character error rate (CER) is a common metric of the performance of an automatic speech recognition system. CER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information. Character error rate can be computed as: CER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct characters, N is the number of characters in the reference (N=S+D+C). CER's output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the performance of the ASR system with a CER of 0 being a perfect score. """ __a = """ Computes CER score of transcribed segments against references. Args: references: list of references for each speech input. predictions: list of transcribtions to score. concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result. Returns: (float): the character error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> cer = datasets.load_metric(\"cer\") >>> cer_score = cer.compute(predictions=predictions, references=references) >>> print(cer_score) 0.34146341463414637 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : Tuple ) -> int: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/jitsi/jiwer/"""] , reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", """https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates""", ] , ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : Optional[Any]=False ) -> List[str]: """simple docstring""" if concatenate_texts: return jiwer.compute_measures( __lowerCamelCase , __lowerCamelCase , truth_transform=__lowerCamelCase , hypothesis_transform=__lowerCamelCase , )["wer"] UpperCAmelCase = 0 UpperCAmelCase = 0 for prediction, reference in zip(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = jiwer.compute_measures( __lowerCamelCase , __lowerCamelCase , truth_transform=__lowerCamelCase , hypothesis_transform=__lowerCamelCase , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
1
import cmath import math def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->complex: UpperCAmelCase = math.radians(lowerCAmelCase_ ) UpperCAmelCase = math.radians(lowerCAmelCase_ ) # Convert voltage and current to rectangular form UpperCAmelCase = cmath.rect(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = cmath.rect(lowerCAmelCase_ , lowerCAmelCase_ ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
627
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
1
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->float: if principal <= 0: raise Exception("""Principal borrowed must be > 0""" ) if rate_per_annum < 0: raise Exception("""Rate of interest must be >= 0""" ) if years_to_repay <= 0 or not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise Exception("""Years to repay must be an integer > 0""" ) # Yearly rate is divided by 12 to get monthly rate UpperCAmelCase = rate_per_annum / 1_2 # Years to repay is multiplied by 12 to get number of payments as payment is monthly UpperCAmelCase = years_to_repay * 1_2 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
627
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
1
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor __a = logging.get_logger(__name__) class __lowercase ( __snake_case ): def __init__( self : Dict , *__lowerCamelCase : Optional[Any] , **__lowerCamelCase : Optional[int] ) -> None: """simple docstring""" warnings.warn( """The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use MobileViTImageProcessor instead.""" , __lowerCamelCase , ) super().__init__(*__lowerCamelCase , **__lowerCamelCase )
627
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[int]: # initialize config if "resnet-50" in model_name: UpperCAmelCase = ResNetConfig.from_pretrained("""microsoft/resnet-50""" ) elif "resnet-101" in model_name: UpperCAmelCase = ResNetConfig.from_pretrained("""microsoft/resnet-101""" ) else: raise ValueError("""Model name should include either resnet50 or resnet101""" ) UpperCAmelCase = DetrConfig(use_timm_backbone=lowerCAmelCase_ , backbone_config=lowerCAmelCase_ ) # set label attributes UpperCAmelCase = """panoptic""" in model_name if is_panoptic: UpperCAmelCase = 2_5_0 else: UpperCAmelCase = 9_1 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """coco-detection-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} return config, is_panoptic def _UpperCamelCase ( lowerCAmelCase_ ) ->Dict: # here we list all keys to be renamed (original name on the left, our name on the right) UpperCAmelCase = [] # stem # fmt: off rename_keys.append(("""backbone.0.body.conv1.weight""", """backbone.conv_encoder.model.embedder.embedder.convolution.weight""") ) rename_keys.append(("""backbone.0.body.bn1.weight""", """backbone.conv_encoder.model.embedder.embedder.normalization.weight""") ) rename_keys.append(("""backbone.0.body.bn1.bias""", """backbone.conv_encoder.model.embedder.embedder.normalization.bias""") ) rename_keys.append(("""backbone.0.body.bn1.running_mean""", """backbone.conv_encoder.model.embedder.embedder.normalization.running_mean""") ) rename_keys.append(("""backbone.0.body.bn1.running_var""", """backbone.conv_encoder.model.embedder.embedder.normalization.running_var""") ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var""", ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean""", ) ) rename_keys.append( ( F"""backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var""", F"""backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var""", ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""", ) ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""") ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""", ) ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""") ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) return rename_keys def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = state_dict.pop(lowerCAmelCase_ ) UpperCAmelCase = val def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_=False ) ->int: UpperCAmelCase = """""" if is_panoptic: UpperCAmelCase = """detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCAmelCase = state_dict.pop(F"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[:2_5_6, :] UpperCAmelCase = in_proj_bias[:2_5_6] UpperCAmelCase = in_proj_weight[2_5_6:5_1_2, :] UpperCAmelCase = in_proj_bias[2_5_6:5_1_2] UpperCAmelCase = in_proj_weight[-2_5_6:, :] UpperCAmelCase = in_proj_bias[-2_5_6:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention UpperCAmelCase = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCAmelCase = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[:2_5_6, :] UpperCAmelCase = in_proj_bias[:2_5_6] UpperCAmelCase = in_proj_weight[2_5_6:5_1_2, :] UpperCAmelCase = in_proj_bias[2_5_6:5_1_2] UpperCAmelCase = in_proj_weight[-2_5_6:, :] UpperCAmelCase = in_proj_bias[-2_5_6:] # read in weights + bias of input projection layer of cross-attention UpperCAmelCase = state_dict.pop( F"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight""" ) UpperCAmelCase = state_dict.pop(F"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) of cross-attention to the state dict UpperCAmelCase = in_proj_weight_cross_attn[:2_5_6, :] UpperCAmelCase = in_proj_bias_cross_attn[:2_5_6] UpperCAmelCase = in_proj_weight_cross_attn[2_5_6:5_1_2, :] UpperCAmelCase = in_proj_bias_cross_attn[2_5_6:5_1_2] UpperCAmelCase = in_proj_weight_cross_attn[-2_5_6:, :] UpperCAmelCase = in_proj_bias_cross_attn[-2_5_6:] def _UpperCamelCase ( ) ->Optional[int]: UpperCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=False ) ->List[Any]: UpperCAmelCase , UpperCAmelCase = get_detr_config(lowerCAmelCase_ ) # load original model from torch hub UpperCAmelCase = { """detr-resnet-50""": """detr_resnet50""", """detr-resnet-101""": """detr_resnet101""", } logger.info(F"""Converting model {model_name}...""" ) UpperCAmelCase = torch.hub.load("""facebookresearch/detr""" , model_name_to_original_name[model_name] , pretrained=lowerCAmelCase_ ).eval() UpperCAmelCase = detr.state_dict() # rename keys for src, dest in create_rename_keys(lowerCAmelCase_ ): if is_panoptic: UpperCAmelCase = """detr.""" + src rename_key(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCAmelCase_ , is_panoptic=lowerCAmelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them UpperCAmelCase = """detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): UpperCAmelCase = state_dict.pop(lowerCAmelCase_ ) UpperCAmelCase = val elif "class_labels_classifier" in key or "bbox_predictor" in key: UpperCAmelCase = state_dict.pop(lowerCAmelCase_ ) UpperCAmelCase = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: UpperCAmelCase = state_dict.pop(lowerCAmelCase_ ) UpperCAmelCase = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): UpperCAmelCase = state_dict.pop(lowerCAmelCase_ ) UpperCAmelCase = val # finally, create HuggingFace model and load state dict UpperCAmelCase = DetrForSegmentation(lowerCAmelCase_ ) if is_panoptic else DetrForObjectDetection(lowerCAmelCase_ ) model.load_state_dict(lowerCAmelCase_ ) model.eval() # verify our conversion on an image UpperCAmelCase = """coco_panoptic""" if is_panoptic else """coco_detection""" UpperCAmelCase = DetrImageProcessor(format=lowerCAmelCase_ ) UpperCAmelCase = processor(images=prepare_img() , return_tensors="""pt""" ) UpperCAmelCase = encoding["""pixel_values"""] UpperCAmelCase = detr(lowerCAmelCase_ ) UpperCAmelCase = model(lowerCAmelCase_ ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1e-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1e-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) processor.save_pretrained(lowerCAmelCase_ ) if push_to_hub: # Upload model and image processor to the hub logger.info("""Uploading PyTorch model and image processor to the hub...""" ) model.push_to_hub(F"""nielsr/{model_name}""" ) processor.push_to_hub(F"""nielsr/{model_name}""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument( """--model_name""", default="""detr-resnet-50""", type=str, choices=["""detr-resnet-50""", """detr-resnet-101"""], help="""Name of the DETR model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Whether to push the model to the hub or not.""") __a = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
1
from __future__ import annotations __a = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->tuple[list[list[int]], list[list[int]]]: UpperCAmelCase = [ [0 for col in range(len(grid[0] ) )] for row in range(len(lowerCAmelCase_ ) ) ] # the reference grid UpperCAmelCase = 1 UpperCAmelCase = [ [0 for col in range(len(grid[0] ) )] for row in range(len(lowerCAmelCase_ ) ) ] # the action grid UpperCAmelCase = init[0] UpperCAmelCase = init[1] UpperCAmelCase = 0 UpperCAmelCase = g + heuristic[x][y] # cost from starting cell to destination cell UpperCAmelCase = [[f, g, x, y]] UpperCAmelCase = False # flag that is set when search is complete UpperCAmelCase = False # flag set if we can't find expand while not found and not resign: if len(lowerCAmelCase_ ) == 0: raise ValueError("""Algorithm is unable to find solution""" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() UpperCAmelCase = cell.pop() UpperCAmelCase = next_cell[2] UpperCAmelCase = next_cell[3] UpperCAmelCase = next_cell[1] if x == goal[0] and y == goal[1]: UpperCAmelCase = True else: for i in range(len(lowerCAmelCase_ ) ): # to try out different valid actions UpperCAmelCase = x + DIRECTIONS[i][0] UpperCAmelCase = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(lowerCAmelCase_ ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: UpperCAmelCase = g + cost UpperCAmelCase = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) UpperCAmelCase = 1 UpperCAmelCase = i UpperCAmelCase = [] UpperCAmelCase = goal[0] UpperCAmelCase = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: UpperCAmelCase = x - DIRECTIONS[action[x][y]][0] UpperCAmelCase = y - DIRECTIONS[action[x][y]][1] UpperCAmelCase = xa UpperCAmelCase = ya invpath.append([x, y] ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): path.append(invpath[len(lowerCAmelCase_ ) - 1 - i] ) return path, action if __name__ == "__main__": __a = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] __a = [0, 0] # all coordinates are given in format [y,x] __a = [len(grid) - 1, len(grid[0]) - 1] __a = 1 # the cost map which pushes the path closer to the goal __a = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): __a = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map __a = 99 __a , __a = search(grid, init, goal, cost, heuristic) print("""ACTION MAP""") for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
627
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def _UpperCamelCase ( lowerCAmelCase_=None ) ->List[str]: UpperCAmelCase = argparse.ArgumentParser(add_help=lowerCAmelCase_ , allow_abbrev=lowerCAmelCase_ ) # The main config parser UpperCAmelCase = config_command_parser(lowerCAmelCase_ ) # The subparser to add commands to UpperCAmelCase = config_parser.add_subparsers(title="""subcommands""" , dest="""subcommand""" ) # Then add other parsers with the parent parser default_command_parser(lowerCAmelCase_ , parents=[parent_parser] ) update_command_parser(lowerCAmelCase_ , parents=[parent_parser] ) return config_parser def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = get_config_parser() UpperCAmelCase = config_parser.parse_args() if not hasattr(lowerCAmelCase_ , """func""" ): config_parser.print_help() exit(1 ) # Run args.func(lowerCAmelCase_ ) if __name__ == "__main__": main()
627
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
1
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( """kwargs, expected""" , [ ({"""num_shards""": 0, """max_num_jobs""": 1}, []), ({"""num_shards""": 1_0, """max_num_jobs""": 1}, [range(1_0 )]), ({"""num_shards""": 1_0, """max_num_jobs""": 1_0}, [range(lowerCAmelCase_ , i + 1 ) for i in range(1_0 )]), ({"""num_shards""": 1, """max_num_jobs""": 1_0}, [range(1 )]), ({"""num_shards""": 1_0, """max_num_jobs""": 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 1_0 )]), ({"""num_shards""": 3, """max_num_jobs""": 1_0}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Any: UpperCAmelCase = _distribute_shards(**lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, max_num_jobs, expected""" , [ ({"""foo""": 0}, 1_0, [{"""foo""": 0}]), ({"""shards""": [0, 1, 2, 3]}, 1, [{"""shards""": [0, 1, 2, 3]}]), ({"""shards""": [0, 1, 2, 3]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}, {"""shards""": [2]}, {"""shards""": [3]}]), ({"""shards""": [0, 1]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}]), ({"""shards""": [0, 1, 2, 3]}, 2, [{"""shards""": [0, 1]}, {"""shards""": [2, 3]}]), ] , ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: UpperCAmelCase = _split_gen_kwargs(lowerCAmelCase_ , lowerCAmelCase_ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, expected""" , [ ({"""foo""": 0}, 1), ({"""shards""": [0]}, 1), ({"""shards""": [0, 1, 2, 3]}, 4), ({"""shards""": [0, 1, 2, 3], """foo""": 0}, 4), ({"""shards""": [0, 1, 2, 3], """other""": (0, 1)}, 4), ({"""shards""": [0, 1, 2, 3], """shards2""": [0, 1]}, RuntimeError), ] , ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Optional[int]: if expected is RuntimeError: with pytest.raises(lowerCAmelCase_ ): _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) else: UpperCAmelCase = _number_of_shards_in_gen_kwargs(lowerCAmelCase_ ) assert out == expected
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
1
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class __lowercase ( unittest.TestCase ): def __init__( self : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple=7 , __lowerCamelCase : Dict=3 , __lowerCamelCase : List[Any]=1_8 , __lowerCamelCase : Union[str, Any]=3_0 , __lowerCamelCase : Optional[int]=4_0_0 , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Union[str, Any]=None , __lowerCamelCase : Any=True , ) -> str: """simple docstring""" UpperCAmelCase = size if size is not None else {"""height""": 1_8, """width""": 1_8} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = image_size UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_normalize def _lowercase ( self : Dict ) -> int: """simple docstring""" return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8_866_443_634_033_203, 0.6_618_829_369_544_983, 0.3_891_746_401_786_804], [-0.6_042_559_146_881_104, -0.02_295_008_860_528_469, 0.5_423_797_369_003_296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = ImageGPTImageProcessor if is_vision_available() else None def _lowercase ( self : Tuple ) -> Optional[int]: """simple docstring""" UpperCAmelCase = ImageGPTImageProcessingTester(self ) @property def _lowercase ( self : Tuple ) -> Optional[Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , """clusters""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_normalize""" ) ) def _lowercase ( self : Optional[int] ) -> int: """simple docstring""" UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 1_8, """width""": 1_8} ) UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 ) self.assertEqual(image_processor.size , {"""height""": 4_2, """width""": 4_2} ) def _lowercase ( self : Dict ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) UpperCAmelCase = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(__lowerCamelCase , obj[key] ) ) else: self.assertEqual(obj[key] , __lowerCamelCase ) def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase = os.path.join(__lowerCamelCase , """image_processor.json""" ) image_processor_first.to_json_file(__lowerCamelCase ) UpperCAmelCase = self.image_processing_class.from_json_file(__lowerCamelCase ).to_dict() UpperCAmelCase = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__lowerCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __lowerCamelCase ) def _lowercase ( self : str ) -> Any: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(__lowerCamelCase ) UpperCAmelCase = self.image_processing_class.from_pretrained(__lowerCamelCase ).to_dict() UpperCAmelCase = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__lowerCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __lowerCamelCase ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" pass def _UpperCamelCase ( ) ->Optional[Any]: UpperCAmelCase = load_dataset("""hf-internal-testing/fixtures_image_utils""" , split="""test""" ) UpperCAmelCase = Image.open(dataset[4]["""file"""] ) UpperCAmelCase = Image.open(dataset[5]["""file"""] ) UpperCAmelCase = [imagea, imagea] return images @require_vision @require_torch class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) UpperCAmelCase = prepare_images() # test non-batched UpperCAmelCase = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1_0_2_4) ) UpperCAmelCase = [3_0_6, 1_9_1, 1_9_1] self.assertEqual(encoding.input_ids[0, :3].tolist() , __lowerCamelCase ) # test batched UpperCAmelCase = image_processing(__lowerCamelCase , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1_0_2_4) ) UpperCAmelCase = [3_0_3, 1_3, 1_3] self.assertEqual(encoding.input_ids[1, -3:].tolist() , __lowerCamelCase )
627
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
1
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
1
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
1
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy __a = logging.getLogger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = False , ) ->Optional[int]: UpperCAmelCase = bnb_quantization_config.load_in_abit UpperCAmelCase = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( """You have a version of `bitsandbytes` that is not compatible with 8bit quantization,""" """ make sure you have the latest version of `bitsandbytes` installed.""" ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( """You have a version of `bitsandbytes` that is not compatible with 4bit quantization,""" """make sure you have the latest version of `bitsandbytes` installed.""" ) UpperCAmelCase = [] # custom device map if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and len(device_map.keys() ) > 1: UpperCAmelCase = [key for key, value in device_map.items() if value in ["""disk""", """cpu"""]] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: UpperCAmelCase = get_keys_to_not_convert(lowerCAmelCase_ ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(lowerCAmelCase_ ) UpperCAmelCase = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: UpperCAmelCase = [] UpperCAmelCase = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(lowerCAmelCase_ ) # compatibility with peft UpperCAmelCase = load_in_abit UpperCAmelCase = load_in_abit UpperCAmelCase = get_parameter_device(lowerCAmelCase_ ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( """It is not recommended to quantize a loaded model. """ """The model should be instantiated under the `init_empty_weights` context manager.""" ) UpperCAmelCase = replace_with_bnb_layers(lowerCAmelCase_ , lowerCAmelCase_ , modules_to_not_convert=lowerCAmelCase_ ) # convert param to the right dtype UpperCAmelCase = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: UpperCAmelCase = name.replace(""".weight""" , """""" ).replace(""".bias""" , """""" ) UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(lowerCAmelCase_ ): param.to(lowerCAmelCase_ ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError("""No GPU found. A GPU is needed for quantization.""" ) logger.info( F"""The model device type is {model_device.type}. However, cuda is needed for quantization.""" """We move the model to cuda.""" ) return model elif weights_location is None: raise RuntimeError( F"""`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} """ ) else: with init_empty_weights(): UpperCAmelCase = replace_with_bnb_layers( lowerCAmelCase_ , lowerCAmelCase_ , modules_to_not_convert=lowerCAmelCase_ ) UpperCAmelCase = get_quantized_model_device_map( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , max_memory=lowerCAmelCase_ , no_split_module_classes=lowerCAmelCase_ , ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): UpperCAmelCase = True UpperCAmelCase = any(x in list(device_map.values() ) for x in ["""cpu""", """disk"""] ) load_checkpoint_in_model( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , dtype=bnb_quantization_config.torch_dtype , offload_folder=lowerCAmelCase_ , offload_state_dict=lowerCAmelCase_ , keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules , offload_abit_bnb=load_in_abit and offload , ) return dispatch_model(lowerCAmelCase_ , device_map=lowerCAmelCase_ , offload_dir=lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_=None ) ->Tuple: if device_map is None: if torch.cuda.is_available(): UpperCAmelCase = {"""""": torch.cuda.current_device()} else: raise RuntimeError("""No GPU found. A GPU is needed for quantization.""" ) logger.info("""The device_map was not initialized.""" """Setting device_map to `{'':torch.cuda.current_device()}`.""" ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( """If passing a string for `device_map`, please choose 'auto', 'balanced', 'balanced_low_0' or """ """'sequential'.""" ) UpperCAmelCase = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) UpperCAmelCase = {} UpperCAmelCase = special_dtypes UpperCAmelCase = no_split_module_classes UpperCAmelCase = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": UpperCAmelCase = get_balanced_memory( lowerCAmelCase_ , low_zero=(device_map == """balanced_low_0""") , max_memory=lowerCAmelCase_ , **lowerCAmelCase_ , ) UpperCAmelCase = max_memory UpperCAmelCase = infer_auto_device_map(lowerCAmelCase_ , **lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): # check if don't have any quantized module on the cpu UpperCAmelCase = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules UpperCAmelCase = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( """ Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit the quantized model. If you want to dispatch the model on the CPU or the disk while keeping these modules in `torch_dtype`, you need to pass a custom `device_map` to `load_and_quantize_model`. Check https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk for more details. """ ) else: logger.info( """Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit""" ) del device_map_without_some_modules return device_map def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None ) ->Optional[int]: if modules_to_not_convert is None: UpperCAmelCase = [] UpperCAmelCase , UpperCAmelCase = _replace_with_bnb_layers( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) if not has_been_replaced: logger.warning( """You are loading your model in 8bit or 4bit but no linear modules were found in your model.""" """ this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers.""" """ Please double check your model architecture, or submit an issue on github if you think this is""" """ a bug.""" ) return model def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None , ) ->Optional[int]: UpperCAmelCase = False for name, module in model.named_children(): if current_key_name is None: UpperCAmelCase = [] current_key_name.append(lowerCAmelCase_ ) if isinstance(lowerCAmelCase_ , nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` UpperCAmelCase = """.""".join(lowerCAmelCase_ ) UpperCAmelCase = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: UpperCAmelCase = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: UpperCAmelCase = bnb.nn.LinearabitLt( module.in_features , module.out_features , module.bias is not None , has_fpaa_weights=lowerCAmelCase_ , threshold=bnb_quantization_config.llm_inta_threshold , ) elif bnb_quantization_config.load_in_abit: UpperCAmelCase = bnb.nn.Linearabit( module.in_features , module.out_features , module.bias is not None , bnb_quantization_config.bnb_abit_compute_dtype , compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant , quant_type=bnb_quantization_config.bnb_abit_quant_type , ) else: raise ValueError("""load_in_8bit and load_in_4bit can't be both False""" ) UpperCAmelCase = module.weight.data if module.bias is not None: UpperCAmelCase = module.bias.data bnb_module.requires_grad_(lowerCAmelCase_ ) setattr(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = True if len(list(module.children() ) ) > 0: UpperCAmelCase , UpperCAmelCase = _replace_with_bnb_layers( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: # Create a copy of the model with init_empty_weights(): UpperCAmelCase = deepcopy(lowerCAmelCase_ ) # this has 0 cost since it is done inside `init_empty_weights` context manager` UpperCAmelCase = find_tied_parameters(lowerCAmelCase_ ) # For compatibility with Accelerate < 0.18 if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: UpperCAmelCase = sum(lowerCAmelCase_ , [] ) UpperCAmelCase = len(lowerCAmelCase_ ) > 0 # Check if it is a base model UpperCAmelCase = False if hasattr(lowerCAmelCase_ , """base_model_prefix""" ): UpperCAmelCase = not hasattr(lowerCAmelCase_ , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head UpperCAmelCase = list(model.named_children() ) UpperCAmelCase = [list_modules[-1][0]] # add last module together with tied weights UpperCAmelCase = set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) UpperCAmelCase = list(set(lowerCAmelCase_ ) ) + list(lowerCAmelCase_ ) # remove ".weight" from the keys UpperCAmelCase = [""".weight""", """.bias"""] UpperCAmelCase = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: UpperCAmelCase = name.replace(lowerCAmelCase_ , """""" ) filtered_module_names.append(lowerCAmelCase_ ) return filtered_module_names def _UpperCamelCase ( lowerCAmelCase_ ) ->List[str]: for m in model.modules(): if isinstance(lowerCAmelCase_ , bnb.nn.Linearabit ): return True return False def _UpperCamelCase ( lowerCAmelCase_ ) ->str: return next(parameter.parameters() ).device def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Any: # if it is not quantized, we quantize and offload the quantized weights and the SCB stats if fpaa_statistics is None: set_module_tensor_to_device(lowerCAmelCase_ , lowerCAmelCase_ , 0 , dtype=lowerCAmelCase_ , value=lowerCAmelCase_ ) UpperCAmelCase = param_name UpperCAmelCase = model if "." in tensor_name: UpperCAmelCase = tensor_name.split(""".""" ) for split in splits[:-1]: UpperCAmelCase = getattr(lowerCAmelCase_ , lowerCAmelCase_ ) if new_module is None: raise ValueError(F"""{module} has no attribute {split}.""" ) UpperCAmelCase = new_module UpperCAmelCase = splits[-1] # offload weights UpperCAmelCase = False offload_weight(module._parameters[tensor_name] , lowerCAmelCase_ , lowerCAmelCase_ , index=lowerCAmelCase_ ) if hasattr(module._parameters[tensor_name] , """SCB""" ): offload_weight( module._parameters[tensor_name].SCB , param_name.replace("""weight""" , """SCB""" ) , lowerCAmelCase_ , index=lowerCAmelCase_ , ) else: offload_weight(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , index=lowerCAmelCase_ ) offload_weight(lowerCAmelCase_ , param_name.replace("""weight""" , """SCB""" ) , lowerCAmelCase_ , index=lowerCAmelCase_ ) set_module_tensor_to_device(lowerCAmelCase_ , lowerCAmelCase_ , """meta""" , dtype=lowerCAmelCase_ , value=torch.empty(*param.size() ) )
627
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
1
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
1
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy __a = logging.get_logger(__name__) class __lowercase ( __snake_case ): def __init__( self : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float , **__lowerCamelCase : Optional[int] ) -> int: """simple docstring""" UpperCAmelCase = feature_size UpperCAmelCase = sampling_rate UpperCAmelCase = padding_value UpperCAmelCase = kwargs.pop("""padding_side""" , """right""" ) UpperCAmelCase = kwargs.pop("""return_attention_mask""" , __lowerCamelCase ) super().__init__(**__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] , __lowerCamelCase : Union[bool, str, PaddingStrategy] = True , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[Union[str, TensorType]] = None , ) -> BatchFeature: """simple docstring""" if isinstance(__lowerCamelCase , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCAmelCase = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( """You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`""" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) UpperCAmelCase = processed_features[self.model_input_names[0]] UpperCAmelCase = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(__lowerCamelCase ) == 0: if return_attention_mask: UpperCAmelCase = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCAmelCase = required_input[0] if isinstance(__lowerCamelCase , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCAmelCase = 0 while len(required_input[index] ) == 0: index += 1 if index < len(__lowerCamelCase ): UpperCAmelCase = required_input[index][0] if return_tensors is None: if is_tf_tensor(__lowerCamelCase ): UpperCAmelCase = """tf""" elif is_torch_tensor(__lowerCamelCase ): UpperCAmelCase = """pt""" elif isinstance(__lowerCamelCase , (int, float, list, tuple, np.ndarray) ): UpperCAmelCase = """np""" else: raise ValueError( F"""type of {first_element} unknown: {type(__lowerCamelCase )}. """ """Should be one of a python, numpy, pytorch or tensorflow object.""" ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCAmelCase = to_numpy(__lowerCamelCase ) else: UpperCAmelCase = [to_numpy(__lowerCamelCase ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCAmelCase = self._get_padding_strategies(padding=__lowerCamelCase , max_length=__lowerCamelCase ) UpperCAmelCase = processed_features[self.model_input_names[0]] UpperCAmelCase = len(__lowerCamelCase ) if not all(len(__lowerCamelCase ) == batch_size for v in processed_features.values() ): raise ValueError("""Some items in the output dictionary have a different batch size than others.""" ) UpperCAmelCase = [] for i in range(__lowerCamelCase ): UpperCAmelCase = {k: v[i] for k, v in processed_features.items()} # truncation UpperCAmelCase = self._truncate( __lowerCamelCase , max_length=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , truncation=__lowerCamelCase , ) truncated_inputs.append(__lowerCamelCase ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCAmelCase = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCAmelCase = PaddingStrategy.MAX_LENGTH UpperCAmelCase = {} for i in range(__lowerCamelCase ): # padding UpperCAmelCase = self._pad( truncated_inputs[i] , max_length=__lowerCamelCase , padding_strategy=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , return_attention_mask=__lowerCamelCase , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCAmelCase = [] if value.dtype is np.dtype(np.floataa ): UpperCAmelCase = value.astype(np.floataa ) batch_outputs[key].append(__lowerCamelCase ) return BatchFeature(__lowerCamelCase , tensor_type=__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Union[Dict[str, np.ndarray], BatchFeature] , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[bool] = None , ) -> dict: """simple docstring""" UpperCAmelCase = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCAmelCase = len(__lowerCamelCase ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(__lowerCamelCase ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCAmelCase = np.ones(len(__lowerCamelCase ) , dtype=np.intaa ) if needs_to_be_padded: UpperCAmelCase = max_length - len(__lowerCamelCase ) if self.padding_side == "right": if return_attention_mask: UpperCAmelCase = np.pad( processed_features["""attention_mask"""] , (0, difference) ) UpperCAmelCase = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCAmelCase = np.pad( __lowerCamelCase , __lowerCamelCase , """constant""" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCAmelCase = np.pad( processed_features["""attention_mask"""] , (difference, 0) ) UpperCAmelCase = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCAmelCase = np.pad( __lowerCamelCase , __lowerCamelCase , """constant""" , constant_values=self.padding_value ) else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return processed_features def _lowercase ( self : int , __lowerCamelCase : Union[Dict[str, np.ndarray], BatchFeature] , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[bool] = None , ) -> Dict: """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("""When setting ``truncation=True``, make sure that ``max_length`` is defined.""" ) UpperCAmelCase = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase = len(__lowerCamelCase ) > max_length if needs_to_be_truncated: UpperCAmelCase = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCAmelCase = processed_features["""attention_mask"""][:max_length] return processed_features def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Any=False , __lowerCamelCase : Any=None ) -> Dict: """simple docstring""" if padding is not False: if padding is True: UpperCAmelCase = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = PaddingStrategy(__lowerCamelCase ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): UpperCAmelCase = padding else: UpperCAmelCase = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( """Asking to pad but the feature_extractor does not have a padding value. Please select a value to use""" """ as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.""" ) return padding_strategy
627
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
1
from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: UpperCAmelCase = [] for part_id in partition_order: UpperCAmelCase = df.where(F"""SPARK_PARTITION_ID() = {part_id}""" ).collect() for row_idx, row in enumerate(lowerCAmelCase_ ): expected_row_ids_and_row_dicts.append((F"""{part_id}_{row_idx}""", row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def _UpperCamelCase ( ) ->List[Any]: UpperCAmelCase = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() UpperCAmelCase = spark.range(1_0_0 ).repartition(1 ) UpperCAmelCase = Spark(lowerCAmelCase_ ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=1_6 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 5_0 @require_not_windows @require_dill_gt_0_3_2 def _UpperCamelCase ( ) ->List[Any]: UpperCAmelCase = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() UpperCAmelCase = spark.range(1_0 ).repartition(2 ) UpperCAmelCase = [1, 0] UpperCAmelCase = _generate_iterable_examples(lowerCAmelCase_ , lowerCAmelCase_ ) # Reverse the partitions. UpperCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCAmelCase_ , lowerCAmelCase_ ) for i, (row_id, row_dict) in enumerate(generate_fn() ): UpperCAmelCase , UpperCAmelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def _UpperCamelCase ( ) ->List[str]: UpperCAmelCase = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() UpperCAmelCase = spark.range(1_0 ).repartition(1 ) UpperCAmelCase = SparkExamplesIterable(lowerCAmelCase_ ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(lowerCAmelCase_ ): assert row_id == F"""0_{i}""" assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def _UpperCamelCase ( ) ->Any: UpperCAmelCase = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() UpperCAmelCase = spark.range(3_0 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch("""numpy.random.Generator""" ) as generator_mock: UpperCAmelCase = lambda lowerCAmelCase_ : x.reverse() UpperCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCAmelCase_ , [2, 1, 0] ) UpperCAmelCase = SparkExamplesIterable(lowerCAmelCase_ ).shuffle_data_sources(lowerCAmelCase_ ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(lowerCAmelCase_ ): UpperCAmelCase , UpperCAmelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() UpperCAmelCase = spark.range(2_0 ).repartition(4 ) # Partitions 0 and 2 UpperCAmelCase = SparkExamplesIterable(lowerCAmelCase_ ).shard_data_sources(worker_id=0 , num_workers=2 ) assert shard_it_a.n_shards == 2 UpperCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCAmelCase_ , [0, 2] ) for i, (row_id, row_dict) in enumerate(lowerCAmelCase_ ): UpperCAmelCase , UpperCAmelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 UpperCAmelCase = SparkExamplesIterable(lowerCAmelCase_ ).shard_data_sources(worker_id=1 , num_workers=2 ) assert shard_it_a.n_shards == 2 UpperCAmelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(lowerCAmelCase_ , [1, 3] ) for i, (row_id, row_dict) in enumerate(lowerCAmelCase_ ): UpperCAmelCase , UpperCAmelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def _UpperCamelCase ( ) ->Optional[Any]: UpperCAmelCase = pyspark.sql.SparkSession.builder.master("""local[*]""" ).appName("""pyspark""" ).getOrCreate() UpperCAmelCase = spark.range(1_0_0 ).repartition(1 ) UpperCAmelCase = Spark(lowerCAmelCase_ ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 1_0_0
627
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
1
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC __a = parse(importlib.metadata.version("""torch""")) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->str: if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) UpperCAmelCase = STR_OPERATION_TO_FUNC[operation] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = parse(importlib.metadata.version(lowerCAmelCase_ ) ) return operation(lowerCAmelCase_ , parse(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Any: return compare_versions(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )
627
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
1
import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params __a = [ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ["""memory_attention""", """encoder_attn"""], ["""attention""", """attn"""], ["""/""", """."""], [""".LayerNorm.gamma""", """_layer_norm.weight"""], [""".LayerNorm.beta""", """_layer_norm.bias"""], ["""r.layer_""", """r.layers."""], ["""output_proj""", """out_proj"""], ["""ffn.dense_1.""", """fc2."""], ["""ffn.dense.""", """fc1."""], ["""ffn_layer_norm""", """final_layer_norm"""], ["""kernel""", """weight"""], ["""encoder_layer_norm.""", """encoder.layer_norm."""], ["""decoder_layer_norm.""", """decoder.layer_norm."""], ["""embeddings.weights""", """shared.weight"""], ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: for pegasus_name, hf_name in PATTERNS: UpperCAmelCase = k.replace(lowerCAmelCase_ , lowerCAmelCase_ ) return k def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->PegasusForConditionalGeneration: UpperCAmelCase = DEFAULTS.copy() cfg_kwargs.update(lowerCAmelCase_ ) UpperCAmelCase = PegasusConfig(**lowerCAmelCase_ ) UpperCAmelCase = PegasusForConditionalGeneration(lowerCAmelCase_ ) UpperCAmelCase = torch_model.model.state_dict() UpperCAmelCase = {} for k, v in tf_weights.items(): UpperCAmelCase = rename_state_dict_key(lowerCAmelCase_ ) if new_k not in sd: raise ValueError(F"""could not find new key {new_k} in state dict. (converted from {k})""" ) if "dense" in k or "proj" in new_k: UpperCAmelCase = v.T UpperCAmelCase = torch.tensor(lowerCAmelCase_ , dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, F"""{new_k}, {k}, {v.shape}, {sd[new_k].shape}""" # make sure embedding.padding_idx is respected UpperCAmelCase = torch.zeros_like(mapping["""shared.weight"""][cfg.pad_token_id + 1] ) UpperCAmelCase = mapping["""shared.weight"""] UpperCAmelCase = mapping["""shared.weight"""] UpperCAmelCase = {k: torch.zeros_like(lowerCAmelCase_ ) for k, v in sd.items() if k.endswith("""bias""" ) and k not in mapping} mapping.update(**lowerCAmelCase_ ) UpperCAmelCase , UpperCAmelCase = torch_model.model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) UpperCAmelCase = [ k for k in missing if k not in ["""encoder.embed_positions.weight""", """decoder.embed_positions.weight"""] ] assert unexpected_missing == [], F"""no matches found for the following torch keys {unexpected_missing}""" assert extra == [], F"""no matches found for the following tf keys {extra}""" return torch_model def _UpperCamelCase ( lowerCAmelCase_="./ckpt/aeslc/model.ckpt-32000" ) ->Dict: UpperCAmelCase = tf.train.list_variables(lowerCAmelCase_ ) UpperCAmelCase = {} UpperCAmelCase = ["""Adafactor""", """global_step"""] for name, shape in tqdm(lowerCAmelCase_ , desc="""converting tf checkpoint to dict""" ): UpperCAmelCase = any(pat in name for pat in ignore_name ) if skip_key: continue UpperCAmelCase = tf.train.load_variable(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = array return tf_weights def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: # save tokenizer first UpperCAmelCase = Path(lowerCAmelCase_ ).parent.name UpperCAmelCase = task_specific_params[F"""summarization_{dataset}"""]["""max_position_embeddings"""] UpperCAmelCase = PegasusTokenizer.from_pretrained("""sshleifer/pegasus""" , model_max_length=lowerCAmelCase_ ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(lowerCAmelCase_ ) # convert model UpperCAmelCase = get_tf_weights_as_numpy(lowerCAmelCase_ ) UpperCAmelCase = task_specific_params[F"""summarization_{dataset}"""] if dataset == "large": UpperCAmelCase = task_specific_params UpperCAmelCase = convert_pegasus(lowerCAmelCase_ , lowerCAmelCase_ ) torch_model.save_pretrained(lowerCAmelCase_ ) UpperCAmelCase = torch_model.state_dict() sd.pop("""model.decoder.embed_positions.weight""" ) sd.pop("""model.encoder.embed_positions.weight""" ) torch.save(lowerCAmelCase_ , Path(lowerCAmelCase_ ) / """pytorch_model.bin""" ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument("""tf_ckpt_path""", type=str, help="""passed to tf.train.list_variables""") parser.add_argument("""save_dir""", default=None, type=str, help="""Path to the output PyTorch model.""") __a = parser.parse_args() if args.save_dir is None: __a = Path(args.tf_ckpt_path).parent.name __a = os.path.join("""pegasus""", dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
627
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
1
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging __a = logging.get_logger(__name__) __a = { """facebook/bart-large""": """https://huggingface.co/facebook/bart-large/resolve/main/config.json""", # See all BART models at https://huggingface.co/models?filter=bart } class __lowercase ( __snake_case ): UpperCamelCase = '''bart''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , __lowerCamelCase : Optional[int]=5_0_2_6_5 , __lowerCamelCase : Optional[Any]=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : List[str]=4_0_9_6 , __lowerCamelCase : str=1_6 , __lowerCamelCase : Dict=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : str=1_6 , __lowerCamelCase : int=0.0 , __lowerCamelCase : str=0.0 , __lowerCamelCase : Tuple="gelu" , __lowerCamelCase : Tuple=1_0_2_4 , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : Dict=0.0 , __lowerCamelCase : int=0.0 , __lowerCamelCase : Dict=0.02 , __lowerCamelCase : Any=0.0 , __lowerCamelCase : Any=False , __lowerCamelCase : int=True , __lowerCamelCase : Dict=3 , __lowerCamelCase : int=1 , __lowerCamelCase : Union[str, Any]=0 , __lowerCamelCase : Tuple=2 , __lowerCamelCase : int=True , __lowerCamelCase : List[str]=2 , __lowerCamelCase : Any=2 , **__lowerCamelCase : Union[str, Any] , ) -> Tuple: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = classifier_dropout UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=__lowerCamelCase , pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , forced_eos_token_id=__lowerCamelCase , **__lowerCamelCase , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , __lowerCamelCase ): UpperCAmelCase = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" ) class __lowercase ( __snake_case ): @property def _lowercase ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: UpperCAmelCase = {0: """batch"""} UpperCAmelCase = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: UpperCAmelCase = {0: """batch""", 1: """decoder_sequence"""} UpperCAmelCase = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(__lowerCamelCase , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: UpperCAmelCase , UpperCAmelCase = self.num_layers for i in range(__lowerCamelCase ): UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} else: UpperCAmelCase = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def _lowercase ( self : Any ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = super().outputs else: UpperCAmelCase = super(__lowerCamelCase , self ).outputs if self.use_past: UpperCAmelCase , UpperCAmelCase = self.num_layers for i in range(__lowerCamelCase ): UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} UpperCAmelCase = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def _lowercase ( self : Optional[int] , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Generate decoder inputs UpperCAmelCase = seq_length if not self.use_past else 1 UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} UpperCAmelCase = dict(**__lowerCamelCase , **__lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch UpperCAmelCase , UpperCAmelCase = common_inputs["""input_ids"""].shape UpperCAmelCase = common_inputs["""decoder_input_ids"""].shape[1] UpperCAmelCase , UpperCAmelCase = self.num_attention_heads UpperCAmelCase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase = decoder_seq_length + 3 UpperCAmelCase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) UpperCAmelCase = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(__lowerCamelCase , __lowerCamelCase )] , dim=1 ) UpperCAmelCase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered UpperCAmelCase , UpperCAmelCase = self.num_layers UpperCAmelCase = min(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = max(__lowerCamelCase , __lowerCamelCase ) - min_num_layers UpperCAmelCase = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(__lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase ), ) ) # TODO: test this. UpperCAmelCase = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(__lowerCamelCase , __lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase )) ) return common_inputs def _lowercase ( self : List[Any] , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch UpperCAmelCase , UpperCAmelCase = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values UpperCAmelCase = seqlen + 2 UpperCAmelCase , UpperCAmelCase = self.num_layers UpperCAmelCase , UpperCAmelCase = self.num_attention_heads UpperCAmelCase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase = common_inputs["""attention_mask"""].dtype UpperCAmelCase = torch.cat( [common_inputs["""attention_mask"""], torch.ones(__lowerCamelCase , __lowerCamelCase , dtype=__lowerCamelCase )] , dim=1 ) UpperCAmelCase = [ (torch.zeros(__lowerCamelCase ), torch.zeros(__lowerCamelCase )) for _ in range(__lowerCamelCase ) ] return common_inputs def _lowercase ( self : str , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCAmelCase = compute_effective_axis_dimension( __lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase = tokenizer.num_special_tokens_to_add(__lowerCamelCase ) UpperCAmelCase = compute_effective_axis_dimension( __lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size UpperCAmelCase = dict(tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase ) ) return common_inputs def _lowercase ( self : int , __lowerCamelCase : PreTrainedTokenizer , __lowerCamelCase : int = -1 , __lowerCamelCase : int = -1 , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __lowerCamelCase , batch_size=__lowerCamelCase , seq_length=__lowerCamelCase , is_pair=__lowerCamelCase , framework=__lowerCamelCase ) elif self.task == "causal-lm": UpperCAmelCase = self._generate_dummy_inputs_for_causal_lm( __lowerCamelCase , batch_size=__lowerCamelCase , seq_length=__lowerCamelCase , is_pair=__lowerCamelCase , framework=__lowerCamelCase ) else: UpperCAmelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __lowerCamelCase , batch_size=__lowerCamelCase , seq_length=__lowerCamelCase , is_pair=__lowerCamelCase , framework=__lowerCamelCase ) return common_inputs def _lowercase ( self : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase = super()._flatten_past_key_values_(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: UpperCAmelCase = super(__lowerCamelCase , self )._flatten_past_key_values_( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase )
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
def _UpperCamelCase ( lowerCAmelCase_ = 5_0 ) ->int: UpperCAmelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
627
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
1
from string import ascii_uppercase __a = {str(ord(c) - 55): c for c in ascii_uppercase} def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""int() can't convert non-string with explicit base""" ) if num < 0: raise ValueError("""parameter must be positive int""" ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""'str' object cannot be interpreted as an integer""" ) if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise TypeError("""'float' object cannot be interpreted as an integer""" ) if base in (0, 1): raise ValueError("""base must be >= 2""" ) if base > 3_6: raise ValueError("""base must be <= 36""" ) UpperCAmelCase = """""" UpperCAmelCase = 0 UpperCAmelCase = 0 while div != 1: UpperCAmelCase , UpperCAmelCase = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) if base >= 1_1 and 9 < mod < 3_6: UpperCAmelCase = ALPHABET_VALUES[str(lowerCAmelCase_ )] else: UpperCAmelCase = str(lowerCAmelCase_ ) new_value += actual_value UpperCAmelCase = num // base UpperCAmelCase = div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(lowerCAmelCase_ ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(1000): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
627
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
1
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class __lowercase ( unittest.TestCase ): UpperCamelCase = JukeboxTokenizer UpperCamelCase = { '''artist''': '''Zac Brown Band''', '''genres''': '''Country''', '''lyrics''': '''I met a traveller from an antique land, Who said "Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away ''', } @require_torch def _lowercase ( self : Tuple ) -> str: """simple docstring""" import torch UpperCAmelCase = JukeboxTokenizer.from_pretrained("""openai/jukebox-1b-lyrics""" ) UpperCAmelCase = tokenizer(**self.metas )["""input_ids"""] # fmt: off UpperCAmelCase = [ torch.tensor([[ 0, 0, 0, 7_1_6_9, 5_0_7, 9, 7_6, 3_9, 3_1, 4_6, 7_6, 2_7, 7_6, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_6, 3_2, 4_4, 4_1, 3_9, 7_6, 2_7, 4_0, 7_6, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_6, 3_8, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 4_1, 7_6, 4_5, 2_7, 3_5, 3_0, 7_6, 7_1, 2_0, 4_9, 4_1, 7_6, 4_8, 2_7, 4_5, 4_6, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_6, 3_8, 3_1, 3_3, 4_5, 7_6, 4_1, 3_2, 7_6, 4_5, 4_6, 4_1, 4_0, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_9, 4_6, 2_7, 4_0, 3_0, 7_6, 3_5, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 1_4, 3_1, 2_7, 4_4, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_5, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 8, 2_7, 3_8, 3_2, 7_6, 4_5, 4_7, 4_0, 3_7, 7_6, 2_7, 7_6, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_6, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_6, 3_8, 3_5, 3_1, 4_5, 6_4, 7_6, 4_9, 3_4, 4_1, 4_5, 3_1, 7_6, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_6, 3_8, 3_5, 4_2, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_5, 4_0, 3_1, 3_1, 4_4, 7_6, 4_1, 3_2, 7_6, 2_9, 4_1, 3_8, 3_0, 7_6, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_5, 4_6, 4_5, 7_6, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_6, 4_9, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 4_1, 4_5, 3_1, 7_6, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_6, 4_4, 3_1, 2_7, 3_0, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 3_5, 2_9, 3_4, 7_6, 5_1, 3_1, 4_6, 7_6, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_6, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_6, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_4, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_4, 3_1, 2_7, 4_4, 4_6, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_2, 3_1, 3_0, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 4_9, 4_1, 4_4, 3_0, 4_5, 7_6, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_3, 5_1, 7_6, 4_0, 2_7, 3_9, 3_1, 7_6, 3_5, 4_5, 7_6, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_6, 1_1, 3_5, 4_0, 3_3, 7_6, 4_1, 3_2, 7_6, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_2, 4_1, 4_1, 3_7, 7_6, 4_1, 4_0, 7_6, 3_9, 5_1, 7_6, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_6, 5_1, 3_1, 7_6, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_6, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_6, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_6, 1_8, 4_1, 4_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 2_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_5, 3_2, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_6, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_6, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_6, 2_7, 4_0, 3_0, 7_6, 2_8, 2_7, 4_4, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_8, 4_1, 4_0, 3_1, 7_6, 2_7, 4_0, 3_0, 7_6, 3_8, 3_1, 4_8, 3_1, 3_8, 7_6, 4_5, 2_7, 4_0, 3_0, 4_5, 7_6, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_6, 3_2, 2_7, 4_4, 7_6, 2_7, 4_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def _lowercase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" import torch UpperCAmelCase = JukeboxTokenizer.from_pretrained("""openai/jukebox-5b-lyrics""" ) UpperCAmelCase = tokenizer(**self.metas )["""input_ids"""] # fmt: off UpperCAmelCase = [ torch.tensor([[ 0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1, 9, 7_7, 3_9, 3_1, 4_6, 7_7, 2_7, 7_7, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_7, 3_2, 4_4, 4_1, 3_9, 7_7, 2_7, 4_0, 7_7, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_7, 3_8, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 4_1, 7_7, 4_5, 2_7, 3_5, 3_0, 7_7, 7_2, 2_0, 4_9, 4_1, 7_7, 4_8, 2_7, 4_5, 4_6, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_7, 3_8, 3_1, 3_3, 4_5, 7_7, 4_1, 3_2, 7_7, 4_5, 4_6, 4_1, 4_0, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_9, 4_6, 2_7, 4_0, 3_0, 7_7, 3_5, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 1_4, 3_1, 2_7, 4_4, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_5, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 8, 2_7, 3_8, 3_2, 7_7, 4_5, 4_7, 4_0, 3_7, 7_7, 2_7, 7_7, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_7, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_7, 3_8, 3_5, 3_1, 4_5, 6_4, 7_7, 4_9, 3_4, 4_1, 4_5, 3_1, 7_7, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_7, 3_8, 3_5, 4_2, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_5, 4_0, 3_1, 3_1, 4_4, 7_7, 4_1, 3_2, 7_7, 2_9, 4_1, 3_8, 3_0, 7_7, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_5, 4_6, 4_5, 7_7, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_7, 4_9, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 4_1, 4_5, 3_1, 7_7, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_7, 4_4, 3_1, 2_7, 3_0, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 3_5, 2_9, 3_4, 7_7, 5_1, 3_1, 4_6, 7_7, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_7, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_7, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_4, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_4, 3_1, 2_7, 4_4, 4_6, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_2, 3_1, 3_0, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 4_9, 4_1, 4_4, 3_0, 4_5, 7_7, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_3, 5_1, 7_7, 4_0, 2_7, 3_9, 3_1, 7_7, 3_5, 4_5, 7_7, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_7, 1_1, 3_5, 4_0, 3_3, 7_7, 4_1, 3_2, 7_7, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_2, 4_1, 4_1, 3_7, 7_7, 4_1, 4_0, 7_7, 3_9, 5_1, 7_7, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_7, 5_1, 3_1, 7_7, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_7, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_7, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_7, 1_8, 4_1, 4_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 2_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_5, 3_2, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_7, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_7, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_7, 2_7, 4_0, 3_0, 7_7, 2_8, 2_7, 4_4, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_8, 4_1, 4_0, 3_1, 7_7, 2_7, 4_0, 3_0, 7_7, 3_8, 3_1, 4_8, 3_1, 3_8, 7_7, 4_5, 2_7, 4_0, 3_0, 4_5, 7_7, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_7, 3_2, 2_7, 4_4, 7_7, 2_7, 4_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
627
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class __lowercase : def __init__( self : str , __lowerCamelCase : Tuple , __lowerCamelCase : Any=1_3 , __lowerCamelCase : Any=7 , __lowerCamelCase : Dict=True , __lowerCamelCase : Any=True , __lowerCamelCase : Any=True , __lowerCamelCase : int=9_9 , __lowerCamelCase : Tuple=3_2 , __lowerCamelCase : List[str]=5 , __lowerCamelCase : int=4 , __lowerCamelCase : Optional[Any]=3_7 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Optional[Any]=5_1_2 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : Union[str, Any]=2 , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Optional[Any]=None , ) -> Dict: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_labels UpperCAmelCase = num_choices UpperCAmelCase = scope UpperCAmelCase = self.vocab_size - 1 def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) UpperCAmelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def _lowercase ( self : int , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Dict , __lowerCamelCase : Any , __lowerCamelCase : Tuple , *__lowerCamelCase : int ) -> List[Any]: """simple docstring""" UpperCAmelCase = OpenAIGPTModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , head_mask=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Dict , *__lowerCamelCase : Any ) -> Dict: """simple docstring""" UpperCAmelCase = OpenAIGPTLMHeadModel(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : int , __lowerCamelCase : List[Any] , __lowerCamelCase : int , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , *__lowerCamelCase : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = OpenAIGPTDoubleHeadsModel(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : List[str] , __lowerCamelCase : Dict , __lowerCamelCase : Tuple , __lowerCamelCase : str , __lowerCamelCase : int , *__lowerCamelCase : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.num_labels UpperCAmelCase = OpenAIGPTForSequenceClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : Dict ) -> int: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) = config_and_inputs UpperCAmelCase = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask, } return config, inputs_dict @require_torch class __lowercase ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) UpperCamelCase = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly UpperCamelCase = ( { '''feature-extraction''': OpenAIGPTModel, '''text-classification''': OpenAIGPTForSequenceClassification, '''text-generation''': OpenAIGPTLMHeadModel, '''zero-shot''': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def _lowercase ( self : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def _lowercase ( self : Tuple , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : List[str]=False ) -> List[Any]: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=__lowerCamelCase , ) UpperCAmelCase = inputs_dict["""labels"""] UpperCAmelCase = inputs_dict["""labels"""] UpperCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=__lowerCamelCase , ) UpperCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) return inputs_dict def _lowercase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = OpenAIGPTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , n_embd=3_7 ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : List[Any] ) -> int: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> int: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*__lowerCamelCase ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> str: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*__lowerCamelCase ) @slow def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = OpenAIGPTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @require_torch class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : Any ) -> Optional[int]: """simple docstring""" UpperCAmelCase = OpenAIGPTLMHeadModel.from_pretrained("""openai-gpt""" ) model.to(__lowerCamelCase ) UpperCAmelCase = torch.tensor([[4_8_1, 4_7_3_5, 5_4_4]] , dtype=torch.long , device=__lowerCamelCase ) # the president is UpperCAmelCase = [ 4_8_1, 4_7_3_5, 5_4_4, 2_4_6, 9_6_3, 8_7_0, 7_6_2, 2_3_9, 2_4_4, 4_0_4_7_7, 2_4_4, 2_4_9, 7_1_9, 8_8_1, 4_8_7, 5_4_4, 2_4_0, 2_4_4, 6_0_3, 4_8_1, ] # the president is a very good man. " \n " i\'m sure he is, " said the UpperCAmelCase = model.generate(__lowerCamelCase , do_sample=__lowerCamelCase ) self.assertListEqual(output_ids[0].tolist() , __lowerCamelCase )
627
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
1
import random def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: UpperCAmelCase = a[left_index] UpperCAmelCase = left_index + 1 for j in range(left_index + 1 , lowerCAmelCase_ ): if a[j] < pivot: UpperCAmelCase , UpperCAmelCase = a[i], a[j] i += 1 UpperCAmelCase , UpperCAmelCase = a[i - 1], a[left_index] return i - 1 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->Union[str, Any]: if left < right: UpperCAmelCase = random.randint(lowerCAmelCase_ , right - 1 ) UpperCAmelCase , UpperCAmelCase = ( a[left], a[pivot], ) # switches the pivot with the left most bound UpperCAmelCase = partition(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) quick_sort_random( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # recursive quicksort to the left of the pivot point quick_sort_random( lowerCAmelCase_ , pivot_index + 1 , lowerCAmelCase_ ) # recursive quicksort to the right of the pivot point def _UpperCamelCase ( ) ->List[str]: UpperCAmelCase = input("""Enter numbers separated by a comma:\n""" ).strip() UpperCAmelCase = [int(lowerCAmelCase_ ) for item in user_input.split(""",""" )] quick_sort_random(lowerCAmelCase_ , 0 , len(lowerCAmelCase_ ) ) print(lowerCAmelCase_ ) if __name__ == "__main__": main()
627
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/data2vec-vision-base-ft""": ( """https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json""" ), } class __lowercase ( __snake_case ): UpperCamelCase = '''data2vec-vision''' def __init__( self : Tuple , __lowerCamelCase : Optional[Any]=7_6_8 , __lowerCamelCase : Union[str, Any]=1_2 , __lowerCamelCase : int=1_2 , __lowerCamelCase : Optional[Any]=3_0_7_2 , __lowerCamelCase : Optional[Any]="gelu" , __lowerCamelCase : Any=0.0 , __lowerCamelCase : str=0.0 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Any=1e-1_2 , __lowerCamelCase : List[Any]=2_2_4 , __lowerCamelCase : Union[str, Any]=1_6 , __lowerCamelCase : List[Any]=3 , __lowerCamelCase : List[Any]=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : int=0.1 , __lowerCamelCase : Tuple=True , __lowerCamelCase : Dict=[3, 5, 7, 1_1] , __lowerCamelCase : Union[str, Any]=[1, 2, 3, 6] , __lowerCamelCase : Any=True , __lowerCamelCase : str=0.4 , __lowerCamelCase : Tuple=2_5_6 , __lowerCamelCase : Dict=1 , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : List[Any]=2_5_5 , **__lowerCamelCase : Dict , ) -> Dict: """simple docstring""" super().__init__(**__lowerCamelCase ) UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = use_mask_token UpperCAmelCase = use_absolute_position_embeddings UpperCAmelCase = use_relative_position_bias UpperCAmelCase = use_shared_relative_position_bias UpperCAmelCase = layer_scale_init_value UpperCAmelCase = drop_path_rate UpperCAmelCase = use_mean_pooling # decode head attributes (semantic segmentation) UpperCAmelCase = out_indices UpperCAmelCase = pool_scales # auxiliary head attributes (semantic segmentation) UpperCAmelCase = use_auxiliary_head UpperCAmelCase = auxiliary_loss_weight UpperCAmelCase = auxiliary_channels UpperCAmelCase = auxiliary_num_convs UpperCAmelCase = auxiliary_concat_input UpperCAmelCase = semantic_loss_ignore_index class __lowercase ( __snake_case ): UpperCamelCase = version.parse('''1.11''' ) @property def _lowercase ( self : str ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def _lowercase ( self : List[Any] ) -> float: """simple docstring""" return 1e-4
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
1
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def _UpperCamelCase ( ) ->None: print("""Making key files...""" ) make_key_files("""rsa""" , 1_0_2_4 ) print("""Key files generation successful.""" ) def _UpperCamelCase ( lowerCAmelCase_ ) ->tuple[tuple[int, int], tuple[int, int]]: print("""Generating prime p...""" ) UpperCAmelCase = rabinMiller.generate_large_prime(lowerCAmelCase_ ) print("""Generating prime q...""" ) UpperCAmelCase = rabinMiller.generate_large_prime(lowerCAmelCase_ ) UpperCAmelCase = p * q print("""Generating e that is relatively prime to (p - 1) * (q - 1)...""" ) while True: UpperCAmelCase = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(lowerCAmelCase_ , (p - 1) * (q - 1) ) == 1: break print("""Calculating d that is mod inverse of e...""" ) UpperCAmelCase = cryptoMath.find_mod_inverse(lowerCAmelCase_ , (p - 1) * (q - 1) ) UpperCAmelCase = (n, e) UpperCAmelCase = (n, d) return (public_key, private_key) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->None: if os.path.exists(F"""{name}_pubkey.txt""" ) or os.path.exists(F"""{name}_privkey.txt""" ): print("""\nWARNING:""" ) print( F"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" """Use a different name or delete these files and re-run this program.""" ) sys.exit() UpperCAmelCase , UpperCAmelCase = generate_key(lowerCAmelCase_ ) print(F"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(F"""{name}_pubkey.txt""" , """w""" ) as out_file: out_file.write(F"""{key_size},{public_key[0]},{public_key[1]}""" ) print(F"""Writing private key to file {name}_privkey.txt...""" ) with open(F"""{name}_privkey.txt""" , """w""" ) as out_file: out_file.write(F"""{key_size},{private_key[0]},{private_key[1]}""" ) if __name__ == "__main__": main()
627
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
1
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = " " ) ->list: UpperCAmelCase = [] UpperCAmelCase = 0 for index, char in enumerate(lowerCAmelCase_ ): if char == separator: split_words.append(string[last_index:index] ) UpperCAmelCase = index + 1 elif index + 1 == len(lowerCAmelCase_ ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
627
def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->int: return int((input_a, input_a).count(0 ) == 0 ) def _UpperCamelCase ( ) ->None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
627
1
import comet # From: unbabel-comet import torch import datasets __a = datasets.logging.get_logger(__name__) __a = """\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel's Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = \"{COMET}: A Neural Framework for {MT} Evaluation\", author = \"Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon\", booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\", month = nov, year = \"2020\", address = \"Online\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\", pages = \"2685--2702\", } """ __a = """\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA's or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. """ __a = """ COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric('comet') >>> # comet_metric = load_metric('comet', 'wmt20-comet-da') # you can also choose which model to use >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"] >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"] >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results[\"scores\"]]) [0.19, 0.92] """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://unbabel.github.io/COMET/html/index.html""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """sources""": datasets.Value("""string""" , id="""sequence""" ), """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/Unbabel/COMET"""] , reference_urls=[ """https://github.com/Unbabel/COMET""", """https://www.aclweb.org/anthology/2020.emnlp-main.213/""", """http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6""", ] , ) def _lowercase ( self : str , __lowerCamelCase : Union[str, Any] ) -> Optional[Any]: """simple docstring""" if self.config_name == "default": UpperCAmelCase = comet.load_from_checkpoint(comet.download_model("""wmt20-comet-da""" ) ) else: UpperCAmelCase = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def _lowercase ( self : str , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : int=False ) -> List[Any]: """simple docstring""" if gpus is None: UpperCAmelCase = 1 if torch.cuda.is_available() else 0 UpperCAmelCase = {"""src""": sources, """mt""": predictions, """ref""": references} UpperCAmelCase = [dict(zip(__lowerCamelCase , __lowerCamelCase ) ) for t in zip(*data.values() )] UpperCAmelCase , UpperCAmelCase = self.scorer.predict(__lowerCamelCase , gpus=__lowerCamelCase , progress_bar=__lowerCamelCase ) return {"mean_score": mean_score, "scores": scores}
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCAmelCase = {"""unk_token""": """<unk>"""} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCAmelCase = { """do_resize""": True, """size""": 2_0, """do_center_crop""": True, """crop_size""": 1_8, """do_normalize""": True, """image_mean""": [0.48_145_466, 0.4_578_275, 0.40_821_073], """image_std""": [0.26_862_954, 0.26_130_258, 0.27_577_711], } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : List[str] ) -> str: """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , **__lowerCamelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Any ) -> List[str]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] UpperCAmelCase = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCAmelCase = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCAmelCase = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """lower newer""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Any ) -> int: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = [["""cat""", """nasa badge"""], ["""person"""]] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = len(__lowerCamelCase ) UpperCAmelCase = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = """google/owlvit-base-patch32""" UpperCAmelCase = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCAmelCase = ["""cat""", """nasa badge"""] UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = 1_6 UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = [ [4_9_4_0_6, 2_3_6_8, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_9_4_0_6, 6_8_4_1, 1_1_3_0_1, 4_9_4_0_7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
1
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = { """configuration_mctct""": ["""MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MCTCTConfig"""], """feature_extraction_mctct""": ["""MCTCTFeatureExtractor"""], """processing_mctct""": ["""MCTCTProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MCTCTForCTC""", """MCTCTModel""", """MCTCTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
from math import sqrt def _UpperCamelCase ( lowerCAmelCase_ = 1_0_0_0_0_0_0 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowerCAmelCase_ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"""{solution() = }""")
627
1
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("""Googling.....""") __a = """https://www.google.com/search?q=""" + """ """.join(sys.argv[1:]) __a = requests.get(url, headers={"""UserAgent""": UserAgent().random}) # res.raise_for_status() with open("""project1a.html""", """wb""") as out_file: # only for knowing the class for data in res.iter_content(1_0000): out_file.write(data) __a = BeautifulSoup(res.text, """html.parser""") __a = list(soup.select(""".eZt8xd"""))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get("""href""")) else: webbrowser.open(F"""https://google.com{link.get("href")}""")
627
from __future__ import annotations def _UpperCamelCase ( lowerCAmelCase_ ) ->None: create_state_space_tree(lowerCAmelCase_ , [] , 0 , [0 for i in range(len(lowerCAmelCase_ ) )] ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , ) ->None: if index == len(lowerCAmelCase_ ): print(lowerCAmelCase_ ) return for i in range(len(lowerCAmelCase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) UpperCAmelCase = True create_state_space_tree(lowerCAmelCase_ , lowerCAmelCase_ , index + 1 , lowerCAmelCase_ ) current_sequence.pop() UpperCAmelCase = False __a = [3, 1, 2, 4] generate_all_permutations(sequence) __a = ["A", "B", "C"] generate_all_permutations(sequence_a)
627
1
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __lowercase ( __snake_case , __snake_case ): UpperCamelCase = 1 @register_to_config def __init__( self : Optional[Any] , __lowerCamelCase : int = 1_0_0_0 , __lowerCamelCase : Optional[Union[np.ndarray, List[float]]] = None ) -> Union[str, Any]: """simple docstring""" self.set_timesteps(__lowerCamelCase ) # standard deviation of the initial noise distribution UpperCAmelCase = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. UpperCAmelCase = 4 # running values UpperCAmelCase = [] def _lowercase ( self : int , __lowerCamelCase : int , __lowerCamelCase : Union[str, torch.device] = None ) -> int: """simple docstring""" UpperCAmelCase = num_inference_steps UpperCAmelCase = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] UpperCAmelCase = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: UpperCAmelCase = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: UpperCAmelCase = torch.sin(steps * math.pi / 2 ) ** 2 UpperCAmelCase = (1.0 - self.betas**2) ** 0.5 UpperCAmelCase = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] UpperCAmelCase = timesteps.to(__lowerCamelCase ) UpperCAmelCase = [] def _lowercase ( self : Optional[int] , __lowerCamelCase : torch.FloatTensor , __lowerCamelCase : int , __lowerCamelCase : torch.FloatTensor , __lowerCamelCase : bool = True , ) -> Union[SchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( """Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler""" ) UpperCAmelCase = (self.timesteps == timestep).nonzero().item() UpperCAmelCase = timestep_index + 1 UpperCAmelCase = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__lowerCamelCase ) if len(self.ets ) == 1: UpperCAmelCase = self.ets[-1] elif len(self.ets ) == 2: UpperCAmelCase = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: UpperCAmelCase = (2_3 * self.ets[-1] - 1_6 * self.ets[-2] + 5 * self.ets[-3]) / 1_2 else: UpperCAmelCase = (1 / 2_4) * (5_5 * self.ets[-1] - 5_9 * self.ets[-2] + 3_7 * self.ets[-3] - 9 * self.ets[-4]) UpperCAmelCase = self._get_prev_sample(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : torch.FloatTensor , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : str ) -> torch.FloatTensor: """simple docstring""" return sample def _lowercase ( self : Optional[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.alphas[timestep_index] UpperCAmelCase = self.betas[timestep_index] UpperCAmelCase = self.alphas[prev_timestep_index] UpperCAmelCase = self.betas[prev_timestep_index] UpperCAmelCase = (sample - sigma * ets) / max(__lowerCamelCase , 1e-8 ) UpperCAmelCase = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self : Union[str, Any] ) -> List[str]: """simple docstring""" return self.config.num_train_timesteps
627
import numpy class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: """simple docstring""" UpperCAmelCase = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase = numpy.random.rand(3 , 1 ) # Real output values provided. UpperCAmelCase = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase = numpy.zeros(output_array.shape ) def _lowercase ( self : List[str] ) -> numpy.ndarray: """simple docstring""" UpperCAmelCase = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def _lowercase ( self : Optional[Any] ) -> None: """simple docstring""" UpperCAmelCase = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) UpperCAmelCase = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) UpperCAmelCase = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def _lowercase ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): UpperCAmelCase = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F"""Iteration {iteration} Loss: {loss}""" ) def _lowercase ( self : List[str] , __lowerCamelCase : numpy.ndarray ) -> int: """simple docstring""" UpperCAmelCase = input_arr UpperCAmelCase = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) UpperCAmelCase = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( lowerCAmelCase_ ) ->numpy.ndarray: return (value) * (1 - (value)) def _UpperCamelCase ( ) ->int: UpperCAmelCase = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase = TwoHiddenLayerNeuralNetwork( input_array=lowerCAmelCase_ , output_array=lowerCAmelCase_ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=lowerCAmelCase_ , iterations=1_0 , give_loss=lowerCAmelCase_ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
627
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, is_batched, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __a = logging.get_logger(__name__) class __lowercase ( __snake_case ): UpperCamelCase = ['''pixel_values'''] def __init__( self : Any , __lowerCamelCase : bool = True , __lowerCamelCase : Optional[Dict[str, int]] = None , __lowerCamelCase : PILImageResampling = PILImageResampling.BICUBIC , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : Union[int, float] = 1 / 2_5_5 , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : bool = True , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , **__lowerCamelCase : int , ) -> None: """simple docstring""" super().__init__(**__lowerCamelCase ) UpperCAmelCase = size if size is not None else {"""height""": 2_2_4, """width""": 2_2_4} UpperCAmelCase = get_size_dict(__lowerCamelCase ) UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 2_2_4, """width""": 2_2_4} UpperCAmelCase = get_size_dict(__lowerCamelCase , default_to_square=__lowerCamelCase , param_name="""crop_size""" ) UpperCAmelCase = do_resize UpperCAmelCase = do_rescale UpperCAmelCase = do_normalize UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = size UpperCAmelCase = resample UpperCAmelCase = rescale_factor UpperCAmelCase = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN UpperCAmelCase = image_std if image_std is not None else IMAGENET_DEFAULT_STD def _lowercase ( self : Union[str, Any] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Dict[str, int] , __lowerCamelCase : PILImageResampling = PILImageResampling.BILINEAR , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : str , ) -> np.ndarray: """simple docstring""" UpperCAmelCase = get_size_dict(__lowerCamelCase ) if "shortest_edge" in size: UpperCAmelCase = get_resize_output_image_size(__lowerCamelCase , size=size["""shortest_edge"""] , default_to_square=__lowerCamelCase ) # size = get_resize_output_image_size(image, size["shortest_edge"], size["longest_edge"]) elif "height" in size and "width" in size: UpperCAmelCase = (size["""height"""], size["""width"""]) else: raise ValueError(F"""Size must contain 'height' and 'width' keys or 'shortest_edge' key. Got {size.keys()}""" ) return resize(__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Dict[str, int] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Tuple , ) -> np.ndarray: """simple docstring""" UpperCAmelCase = get_size_dict(__lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(__lowerCamelCase , size=(size["""height"""], size["""width"""]) , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : List[Any] , __lowerCamelCase : np.ndarray , __lowerCamelCase : float , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Tuple ) -> np.ndarray: """simple docstring""" return rescale(__lowerCamelCase , scale=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Union[float, List[float]] , __lowerCamelCase : Union[float, List[float]] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : List[str] , ) -> np.ndarray: """simple docstring""" return normalize(__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : ImageInput , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : PILImageResampling = None , __lowerCamelCase : bool = None , __lowerCamelCase : int = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[float] = None , __lowerCamelCase : Optional[bool] = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : Optional[Union[str, TensorType]] = None , __lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **__lowerCamelCase : Optional[int] , ) -> BatchFeature: """simple docstring""" UpperCAmelCase = do_resize if do_resize is not None else self.do_resize UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase = crop_size if crop_size is not None else self.crop_size UpperCAmelCase = get_size_dict(__lowerCamelCase , param_name="""crop_size""" , default_to_square=__lowerCamelCase ) UpperCAmelCase = resample if resample is not None else self.resample UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase = image_mean if image_mean is not None else self.image_mean UpperCAmelCase = image_std if image_std is not None else self.image_std UpperCAmelCase = size if size is not None else self.size UpperCAmelCase = get_size_dict(__lowerCamelCase ) if not is_batched(__lowerCamelCase ): UpperCAmelCase = [images] if not valid_images(__lowerCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) # All transformations expect numpy arrays. UpperCAmelCase = [to_numpy_array(__lowerCamelCase ) for image in images] if do_resize: UpperCAmelCase = [self.resize(image=__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase ) for image in images] if do_center_crop: UpperCAmelCase = [self.center_crop(image=__lowerCamelCase , size=__lowerCamelCase ) for image in images] if do_rescale: UpperCAmelCase = [self.rescale(image=__lowerCamelCase , scale=__lowerCamelCase ) for image in images] if do_normalize: UpperCAmelCase = [self.normalize(image=__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase ) for image in images] UpperCAmelCase = [to_channel_dimension_format(__lowerCamelCase , __lowerCamelCase ) for image in images] UpperCAmelCase = {"""pixel_values""": images} return BatchFeature(data=__lowerCamelCase , tensor_type=__lowerCamelCase )
627
import argparse __a = """docs/source/_static/js/custom.js""" def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: with open(lowerCAmelCase_ , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCAmelCase = f.readlines() UpperCAmelCase = 0 # First let's put the right version while not lines[index].startswith("""const stableVersion =""" ): index += 1 UpperCAmelCase = F"""const stableVersion = \"v{version}\"\n""" # Then update the dictionary while not lines[index].startswith("""const versionMapping = {""" ): index += 1 # We go until the end while not lines[index].startswith("""}""" ): index += 1 # We add the new version at the end lines[index - 1] += F""" \"v{version}\": \"v{version}\",\n""" with open(lowerCAmelCase_ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __a = parser.parse_args() update_custom_js(args.version)
627
1
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __a = logging.get_logger(__name__) __a = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def _UpperCamelCase ( lowerCAmelCase_ ) ->Any: if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: UpperCAmelCase = k.replace(lowerCAmelCase_ , lowerCAmelCase_ ) if k.startswith("""encoder""" ): UpperCAmelCase = k.replace(""".attn""" , """.self_attn""" ) UpperCAmelCase = k.replace("""norm1""" , """self_attn_layer_norm""" ) UpperCAmelCase = k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): UpperCAmelCase = k.replace("""norm1""" , """self_attn_layer_norm""" ) UpperCAmelCase = k.replace("""norm2""" , """encoder_attn_layer_norm""" ) UpperCAmelCase = k.replace("""norm3""" , """final_layer_norm""" ) return k def _UpperCamelCase ( lowerCAmelCase_ ) ->Union[str, Any]: UpperCAmelCase = [ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: UpperCAmelCase = sd.pop(lowerCAmelCase_ ) UpperCAmelCase = k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd UpperCAmelCase = v __a = ["""START"""] @torch.no_grad() def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->int: UpperCAmelCase = torch.load(lowerCAmelCase_ , map_location="""cpu""" ) UpperCAmelCase = model["""model"""] UpperCAmelCase = BlenderbotConfig.from_json_file(lowerCAmelCase_ ) UpperCAmelCase = BlenderbotForConditionalGeneration(lowerCAmelCase_ ) UpperCAmelCase = m.model.state_dict().keys() UpperCAmelCase = [] UpperCAmelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue UpperCAmelCase = rename_state_dict_key(lowerCAmelCase_ ) if new_k not in valid_keys: failures.append([k, new_k] ) else: UpperCAmelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(lowerCAmelCase_ ) m.model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) m.half() m.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) __a = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
627
import math class __lowercase : def _lowercase ( self : Union[str, Any] , __lowerCamelCase : list[list[float]] , __lowerCamelCase : list[int] ) -> int: """simple docstring""" UpperCAmelCase = 0.0 UpperCAmelCase = 0.0 for i in range(len(__lowerCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _lowercase ( self : List[Any] , __lowerCamelCase : list[list[int | float]] , __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : float ) -> list[list[int | float]]: """simple docstring""" for i in range(len(__lowerCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _UpperCamelCase ( ) ->None: # Training Examples ( m, n ) UpperCAmelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase = SelfOrganizingMap() UpperCAmelCase = 3 UpperCAmelCase = 0.5 for _ in range(lowerCAmelCase_ ): for j in range(len(lowerCAmelCase_ ) ): # training sample UpperCAmelCase = training_samples[j] # Compute the winning vector UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # Update the winning vector UpperCAmelCase = self_organizing_map.update(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # classify test sample UpperCAmelCase = [0, 0, 0, 1] UpperCAmelCase = self_organizing_map.get_winner(lowerCAmelCase_ , lowerCAmelCase_ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
627
1
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {} UpperCAmelCase = tokenizer(example["""content"""] , truncation=lowerCAmelCase_ )["""input_ids"""] UpperCAmelCase = len(example["""content"""] ) / len(output["""input_ids"""] ) return output __a = HfArgumentParser(PretokenizationArguments) __a = parser.parse_args() if args.num_workers is None: __a = multiprocessing.cpu_count() __a = AutoTokenizer.from_pretrained(args.tokenizer_dir) __a = time.time() __a = load_dataset(args.dataset_name, split="""train""") print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") __a = time.time() __a = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ """repo_name""", """path""", """copies""", """size""", """content""", """license""", """hash""", """line_mean""", """line_max""", """alpha_frac""", """autogenerated""", ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") __a = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
627
1
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_=None ) ->Union[str, Any]: if attention_mask is None: UpperCAmelCase = tf.cast(tf.math.not_equal(lowerCAmelCase_ , config.pad_token_id ) , tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowercase : UpperCamelCase = OPTConfig UpperCamelCase = {} UpperCamelCase = '''gelu''' def __init__( self : Dict , __lowerCamelCase : Any , __lowerCamelCase : Union[str, Any]=1_3 , __lowerCamelCase : int=7 , __lowerCamelCase : List[str]=True , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Any=9_9 , __lowerCamelCase : Optional[int]=1_6 , __lowerCamelCase : List[str]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : str=4 , __lowerCamelCase : Optional[int]="gelu" , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : Any=2_0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Dict=1 , __lowerCamelCase : List[Any]=0 , __lowerCamelCase : Union[str, Any]=1_6 , __lowerCamelCase : Union[str, Any]=1_6 , ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = eos_token_id UpperCAmelCase = pad_token_id UpperCAmelCase = bos_token_id UpperCAmelCase = embed_dim UpperCAmelCase = word_embed_proj_dim UpperCAmelCase = False def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__lowerCamelCase , **self.config_updates , ) UpperCAmelCase = prepare_opt_inputs_dict(__lowerCamelCase , __lowerCamelCase ) return config, inputs_dict def _lowercase ( self : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : Dict ) -> int: """simple docstring""" UpperCAmelCase = TFOPTModel(config=__lowerCamelCase ) UpperCAmelCase = inputs_dict["""input_ids"""] UpperCAmelCase = input_ids[:1, :] UpperCAmelCase = inputs_dict["""attention_mask"""][:1, :] UpperCAmelCase = 1 # first forward pass UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase , use_cache=__lowerCamelCase ) UpperCAmelCase , UpperCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] UpperCAmelCase = model(__lowerCamelCase , attention_mask=__lowerCamelCase , past_key_values=__lowerCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCAmelCase = output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__lowerCamelCase , __lowerCamelCase , rtol=1e-3 ) @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () UpperCamelCase = (TFOPTForCausalLM,) if is_tf_available() else () UpperCamelCase = ( {'''feature-extraction''': TFOPTModel, '''text-generation''': TFOPTForCausalLM} if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = 10 def _lowercase ( self : str ) -> Dict: """simple docstring""" UpperCAmelCase = TFOPTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase ) def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__lowerCamelCase ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__lowerCamelCase : int , __lowerCamelCase : Tuple ): if hasattr(__lowerCamelCase , """weight""" ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__lowerCamelCase , """weight""" ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 1_0, config.vocab_size + 1_0]: # build the embeddings UpperCAmelCase = model_class(config=__lowerCamelCase ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_input_embeddings() ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__lowerCamelCase ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_input_embeddings() ) UpperCAmelCase = _get_word_embedding_weight(__lowerCamelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. UpperCAmelCase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __lowerCamelCase ) # check that weights remain the same after resizing UpperCAmelCase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase = False self.assertTrue(__lowerCamelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __lowerCamelCase ) UpperCAmelCase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase = False self.assertTrue(__lowerCamelCase ) def _UpperCamelCase ( lowerCAmelCase_ ) ->Optional[Any]: return tf.constant(lowerCAmelCase_ , dtype=tf.intaa ) @require_tf class __lowercase ( unittest.TestCase ): UpperCamelCase = 99 def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 UpperCAmelCase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) UpperCAmelCase = input_ids.shape[0] UpperCAmelCase = OPTConfig( vocab_size=self.vocab_size , hidden_size=2_4 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase = TFOPTModel.from_pretrained("""facebook/opt-350m""" ) UpperCAmelCase = _long_tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) UpperCAmelCase = tf.not_equal(__lowerCamelCase , model.config.pad_token_id ) with tf.GradientTape(): UpperCAmelCase = model(input_ids=__lowerCamelCase , attention_mask=__lowerCamelCase ).last_hidden_state UpperCAmelCase = (1, 1_1, 5_1_2) self.assertEqual(output.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __lowerCamelCase , atol=4e-3 ) ) UpperCAmelCase = tf.function(__lowerCamelCase , jit_compile=__lowerCamelCase ) UpperCAmelCase = xla_generate(__lowerCamelCase , __lowerCamelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __lowerCamelCase , atol=4e-2 ) ) @require_tf @slow class __lowercase ( unittest.TestCase ): def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" super().setUp() UpperCAmelCase = """facebook/opt-350m""" def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase = TFOPTForCausalLM.from_pretrained(self.path_model ) UpperCAmelCase = GPTaTokenizer.from_pretrained(self.path_model ) UpperCAmelCase = [ """Today is a beautiful day and I want to""", """In the city of""", """Paris is the capital of France and""", """Computers and mobile phones have taken""", ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" , padding=__lowerCamelCase , add_special_tokens=__lowerCamelCase ) UpperCAmelCase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) UpperCAmelCase = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-4 ) ) UpperCAmelCase = tf.function(__lowerCamelCase , jit_compile=__lowerCamelCase ) UpperCAmelCase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-4 ) ) @require_tf @slow class __lowercase ( unittest.TestCase ): @property def _lowercase ( self : Any ) -> str: """simple docstring""" return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase = """facebook/opt-125m""" UpperCAmelCase = [ """Today is a beautiful day and I want to""", """In the city of New York, the city""", """Paris is the capital of France and the capital""", """Computers and mobile phones have taken over the""", ] UpperCAmelCase = [] UpperCAmelCase = GPTaTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) for prompt in self.prompts: UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(__lowerCamelCase , max_length=1_0 ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) predicted_outputs += generated_string self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Tuple ) -> str: """simple docstring""" UpperCAmelCase = """facebook/opt-350m""" UpperCAmelCase = GPTaTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) UpperCAmelCase = """left""" # use different length sentences to test batching UpperCAmelCase = [ """Hello, my dog is a little""", """Today, I""", ] UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" , padding=__lowerCamelCase ) UpperCAmelCase = inputs["""input_ids"""] UpperCAmelCase = model.generate(input_ids=__lowerCamelCase , attention_mask=inputs["""attention_mask"""] ) UpperCAmelCase = tokenizer(sentences[0] , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(input_ids=__lowerCamelCase ) UpperCAmelCase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs["""attention_mask"""][-1] , tf.intaa ) ) UpperCAmelCase = tokenizer(sentences[1] , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(input_ids=__lowerCamelCase , max_length=model.config.max_length - num_paddings ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__lowerCamelCase ) UpperCAmelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=__lowerCamelCase ) UpperCAmelCase = [ """Hello, my dog is a little bit of a dork.\nI'm a little bit""", """Today, I was in the middle of a conversation with a friend about the""", ] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(__lowerCamelCase , [non_padded_sentence, padded_sentence] ) def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = """facebook/opt-350m""" UpperCAmelCase = [ """Today is a beautiful day and I want to""", """In the city of San Francisco, the city""", """Paris is the capital of France and the capital""", """Computers and mobile phones have taken over the""", ] UpperCAmelCase = [] UpperCAmelCase = GPTaTokenizer.from_pretrained(__lowerCamelCase ) UpperCAmelCase = TFOPTForCausalLM.from_pretrained(__lowerCamelCase ) for prompt in self.prompts: UpperCAmelCase = tokenizer(__lowerCamelCase , return_tensors="""tf""" ).input_ids UpperCAmelCase = model.generate(__lowerCamelCase , max_length=1_0 ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) predicted_outputs += generated_string self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
627
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __a = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ __a = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ __a = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[str]: return float((preds == labels).mean() ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_="binary" ) ->Union[str, Any]: UpperCAmelCase = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->List[Any]: UpperCAmelCase = {} for id_pred, label in zip(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" UpperCAmelCase = id_pred["""prediction"""] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCAmelCase = [(pred, label)] UpperCAmelCase , UpperCAmelCase = [], [] for question, preds_labels in question_map.items(): UpperCAmelCase , UpperCAmelCase = zip(*lowerCAmelCase_ ) UpperCAmelCase = fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ , average="""macro""" ) fas.append(lowerCAmelCase_ ) UpperCAmelCase = int(sum(pred == label for pred, label in preds_labels ) == len(lowerCAmelCase_ ) ) ems.append(lowerCAmelCase_ ) UpperCAmelCase = float(sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) ) UpperCAmelCase = sum(lowerCAmelCase_ ) / len(lowerCAmelCase_ ) UpperCAmelCase = float(fa_score(y_true=lowerCAmelCase_ , y_pred=[id_pred["""prediction"""] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): def _lowercase ( self : int ) -> Any: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if not self.config_name == """record""" and not self.config_name == """multirc""" else None , ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "prediction_text": datasets.Value("""string""" ), }, "references": { "idx": { "passage": datasets.Value("""int64""" ), "query": datasets.Value("""int64""" ), }, "answers": datasets.Sequence(datasets.Value("""string""" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("""int64""" ), "paragraph": datasets.Value("""int64""" ), "question": datasets.Value("""int64""" ), }, "prediction": datasets.Value("""int64""" ), }, "references": datasets.Value("""int64""" ), } else: return { "predictions": datasets.Value("""int64""" ), "references": datasets.Value("""int64""" ), } def _lowercase ( self : Dict , __lowerCamelCase : int , __lowerCamelCase : Optional[int] ) -> List[Any]: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg="""macro""" ) elif self.config_name == "record": UpperCAmelCase = [ { """qas""": [ {"""id""": ref["""idx"""]["""query"""], """answers""": [{"""text""": ans} for ans in ref["""answers"""]]} for ref in references ] } ] UpperCAmelCase = {pred["""idx"""]["""query"""]: pred["""prediction_text"""] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]""" )
627
1
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class __lowercase ( unittest.TestCase ): def __init__( self : int , __lowerCamelCase : Tuple , __lowerCamelCase : int=1_3 , __lowerCamelCase : Tuple=7 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : List[str]=True , __lowerCamelCase : int=True , __lowerCamelCase : Dict=True , __lowerCamelCase : Dict=9_9 , __lowerCamelCase : Dict=3_2 , __lowerCamelCase : int=5 , __lowerCamelCase : Optional[Any]=4 , __lowerCamelCase : Optional[int]=3_7 , __lowerCamelCase : Tuple="gelu" , __lowerCamelCase : List[Any]=0.1 , __lowerCamelCase : Dict=0.1 , __lowerCamelCase : Optional[int]=5_1_2 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : str=2 , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : Optional[Any]=4 , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = seq_length UpperCAmelCase = is_training UpperCAmelCase = use_attention_mask UpperCAmelCase = use_token_type_ids UpperCAmelCase = use_labels UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = num_choices def _lowercase ( self : int ) -> Dict: """simple docstring""" UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase = None if self.use_attention_mask: UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase = None if self.use_token_type_ids: UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _lowercase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = True UpperCAmelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = True UpperCamelCase = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase = FlaxRobertaModelTester(self ) @slow def _lowercase ( self : Tuple ) -> str: """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase = model_class_name.from_pretrained("""roberta-base""" , from_pt=__lowerCamelCase ) UpperCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowerCamelCase )
627
import math import qiskit def _UpperCamelCase ( lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1 ) ->qiskit.result.counts.Counts: if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) or isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != input_a) or (math.floor(lowerCAmelCase_ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCAmelCase = qiskit.QuantumRegister(4 , """qr""" ) UpperCAmelCase = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCAmelCase = [input_a, input_a, carry_in] UpperCAmelCase = qiskit.QuantumCircuit(lowerCAmelCase_ , lowerCAmelCase_ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(lowerCAmelCase_ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(lowerCAmelCase_ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(lowerCAmelCase_ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , lowerCAmelCase_ ) # measure the last two qbits UpperCAmelCase = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCAmelCase = qiskit.execute(lowerCAmelCase_ , lowerCAmelCase_ , shots=1_0_0_0 ) return job.result().get_counts(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
627
1
import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_="shi-labs/oneformer_demo" ) ->str: with open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type="""dataset""" ) , """r""" ) as f: UpperCAmelCase = json.load(lowerCAmelCase_ ) UpperCAmelCase = {} UpperCAmelCase = [] UpperCAmelCase = [] for key, info in class_info.items(): UpperCAmelCase = info["""name"""] class_names.append(info["""name"""] ) if info["isthing"]: thing_ids.append(int(lowerCAmelCase_ ) ) UpperCAmelCase = thing_ids UpperCAmelCase = class_names return metadata class __lowercase ( unittest.TestCase ): def __init__( self : int , __lowerCamelCase : str , __lowerCamelCase : Tuple=7 , __lowerCamelCase : Optional[Any]=3 , __lowerCamelCase : str=3_0 , __lowerCamelCase : List[str]=4_0_0 , __lowerCamelCase : Any=None , __lowerCamelCase : Tuple=True , __lowerCamelCase : Dict=True , __lowerCamelCase : Tuple=[0.5, 0.5, 0.5] , __lowerCamelCase : Tuple=[0.5, 0.5, 0.5] , __lowerCamelCase : Optional[int]=1_0 , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : int=2_5_5 , __lowerCamelCase : Optional[Any]="shi-labs/oneformer_demo" , __lowerCamelCase : Optional[int]="ade20k_panoptic.json" , __lowerCamelCase : Any=1_0 , ) -> Optional[int]: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = {"""shortest_edge""": 3_2, """longest_edge""": 1_3_3_3} if size is None else size UpperCAmelCase = do_normalize UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = class_info_file UpperCAmelCase = prepare_metadata(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = num_text UpperCAmelCase = repo_path # for the post_process_functions UpperCAmelCase = 2 UpperCAmelCase = 1_0 UpperCAmelCase = 1_0 UpperCAmelCase = 3 UpperCAmelCase = 4 UpperCAmelCase = num_labels UpperCAmelCase = do_reduce_labels UpperCAmelCase = ignore_index def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def _lowercase ( self : List[str] , __lowerCamelCase : int , __lowerCamelCase : int=False ) -> int: """simple docstring""" if not batched: UpperCAmelCase = image_inputs[0] if isinstance(__lowerCamelCase , Image.Image ): UpperCAmelCase , UpperCAmelCase = image.size else: UpperCAmelCase , UpperCAmelCase = image.shape[1], image.shape[2] if w < h: UpperCAmelCase = int(self.size["""shortest_edge"""] * h / w ) UpperCAmelCase = self.size["""shortest_edge"""] elif w > h: UpperCAmelCase = self.size["""shortest_edge"""] UpperCAmelCase = int(self.size["""shortest_edge"""] * w / h ) else: UpperCAmelCase = self.size["""shortest_edge"""] UpperCAmelCase = self.size["""shortest_edge"""] else: UpperCAmelCase = [] for image in image_inputs: UpperCAmelCase , UpperCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCAmelCase = max(__lowerCamelCase , key=lambda __lowerCamelCase : item[0] )[0] UpperCAmelCase = max(__lowerCamelCase , key=lambda __lowerCamelCase : item[1] )[1] return expected_height, expected_width def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ) , masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ) , ) @require_torch @require_vision class __lowercase ( __snake_case , unittest.TestCase ): UpperCamelCase = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string UpperCamelCase = image_processing_class def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = OneFormerImageProcessorTester(self ) @property def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" return self.image_processing_tester.prepare_image_processor_dict() def _lowercase ( self : Optional[Any] ) -> str: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """size""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """ignore_index""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """class_info_file""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """num_text""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """repo_path""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """metadata""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_reduce_labels""" ) ) def _lowercase ( self : Tuple ) -> int: """simple docstring""" pass def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , Image.Image ) # Test not batched input UpperCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processing_tester.get_expected_values(__lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase , UpperCAmelCase = self.image_processing_tester.get_expected_values(__lowerCamelCase , batched=__lowerCamelCase ) UpperCAmelCase = image_processor( __lowerCamelCase , ["""semantic"""] * len(__lowerCamelCase ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=__lowerCamelCase , numpify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , np.ndarray ) # Test not batched input UpperCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processing_tester.get_expected_values(__lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase , UpperCAmelCase = self.image_processing_tester.get_expected_values(__lowerCamelCase , batched=__lowerCamelCase ) UpperCAmelCase = image_processor( __lowerCamelCase , ["""semantic"""] * len(__lowerCamelCase ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def _lowercase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=__lowerCamelCase , torchify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values UpperCAmelCase , UpperCAmelCase = self.image_processing_tester.get_expected_values(__lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCAmelCase , UpperCAmelCase = self.image_processing_tester.get_expected_values(__lowerCamelCase , batched=__lowerCamelCase ) UpperCAmelCase = image_processor( __lowerCamelCase , ["""semantic"""] * len(__lowerCamelCase ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Tuple=False , __lowerCamelCase : int="np" ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # prepare image and target UpperCAmelCase = self.image_processing_tester.num_labels UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=__lowerCamelCase ) if with_segmentation_maps: UpperCAmelCase = num_labels if is_instance_map: UpperCAmelCase = list(range(__lowerCamelCase ) ) * 2 UpperCAmelCase = dict(enumerate(__lowerCamelCase ) ) UpperCAmelCase = [ np.random.randint(0 , high * 2 , (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": UpperCAmelCase = [Image.fromarray(__lowerCamelCase ) for annotation in annotations] UpperCAmelCase = image_processor( __lowerCamelCase , ["""semantic"""] * len(__lowerCamelCase ) , __lowerCamelCase , return_tensors="""pt""" , instance_id_to_semantic_id=__lowerCamelCase , pad_and_return_pixel_mask=__lowerCamelCase , ) return inputs def _lowercase ( self : str ) -> Any: """simple docstring""" pass def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" def common(__lowerCamelCase : Tuple=False , __lowerCamelCase : List[str]=None ): UpperCAmelCase = self.comm_get_image_processor_inputs( with_segmentation_maps=__lowerCamelCase , is_instance_map=__lowerCamelCase , segmentation_type=__lowerCamelCase ) UpperCAmelCase = inputs["""mask_labels"""] UpperCAmelCase = inputs["""class_labels"""] UpperCAmelCase = inputs["""pixel_values"""] UpperCAmelCase = inputs["""text_inputs"""] # check the batch_size for mask_label, class_label, text_input in zip(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): self.assertEqual(mask_label.shape[0] , class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:] , pixel_values.shape[2:] ) self.assertEqual(len(__lowerCamelCase ) , self.image_processing_tester.num_text ) common() common(is_instance_map=__lowerCamelCase ) common(is_instance_map=__lowerCamelCase , segmentation_type="""pil""" ) common(is_instance_map=__lowerCamelCase , segmentation_type="""pil""" ) def _lowercase ( self : List[str] ) -> Dict: """simple docstring""" UpperCAmelCase = np.zeros((2_0, 5_0) ) UpperCAmelCase = 1 UpperCAmelCase = 1 UpperCAmelCase = 1 UpperCAmelCase = binary_mask_to_rle(__lowerCamelCase ) self.assertEqual(len(__lowerCamelCase ) , 4 ) self.assertEqual(rle[0] , 2_1 ) self.assertEqual(rle[1] , 4_5 ) def _lowercase ( self : Dict ) -> Any: """simple docstring""" UpperCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=7_7 , task_seq_length=7_7 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) UpperCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() UpperCAmelCase = fature_extractor.post_process_semantic_segmentation(__lowerCamelCase ) self.assertEqual(len(__lowerCamelCase ) , self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape , ( self.image_processing_tester.height, self.image_processing_tester.width, ) , ) UpperCAmelCase = [(1, 4) for i in range(self.image_processing_tester.batch_size )] UpperCAmelCase = fature_extractor.post_process_semantic_segmentation(__lowerCamelCase , target_sizes=__lowerCamelCase ) self.assertEqual(segmentation[0].shape , target_sizes[0] ) def _lowercase ( self : Any ) -> Any: """simple docstring""" UpperCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=7_7 , task_seq_length=7_7 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) UpperCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() UpperCAmelCase = image_processor.post_process_instance_segmentation(__lowerCamelCase , threshold=0 ) self.assertTrue(len(__lowerCamelCase ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , __lowerCamelCase ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) ) def _lowercase ( self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=7_7 , task_seq_length=7_7 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) UpperCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() UpperCAmelCase = image_processor.post_process_panoptic_segmentation(__lowerCamelCase , threshold=0 ) self.assertTrue(len(__lowerCamelCase ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , __lowerCamelCase ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) )
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class __lowercase ( __snake_case ): UpperCamelCase = '''ctrl''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : str , __lowerCamelCase : Optional[int]=2_4_6_5_3_4 , __lowerCamelCase : Union[str, Any]=2_5_6 , __lowerCamelCase : int=1_2_8_0 , __lowerCamelCase : Optional[Any]=8_1_9_2 , __lowerCamelCase : List[str]=4_8 , __lowerCamelCase : Dict=1_6 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=1e-6 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Union[str, Any]=True , **__lowerCamelCase : List[str] , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = dff UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = use_cache super().__init__(**__lowerCamelCase )
627
1
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __lowercase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCAmelCase = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 1_6_0_0_0, """return_attention_mask""": False, """do_normalize""": True, } UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.feature_extraction_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) # load decoder from hub UpperCAmelCase = """hf-internal-testing/ngram-beam-search-decoder""" def _lowercase ( self : Optional[int] , **__lowerCamelCase : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = self.add_kwargs_tokens_map.copy() kwargs.update(__lowerCamelCase ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : List[str] , **__lowerCamelCase : str ) -> str: """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : List[Any] , **__lowerCamelCase : int ) -> Tuple: """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **__lowerCamelCase ) def _lowercase ( self : Any ) -> str: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __lowerCamelCase ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , __lowerCamelCase ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(__lowerCamelCase , """include""" ): WavaVecaProcessorWithLM( tokenizer=__lowerCamelCase , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = floats_list((3, 1_0_0_0) ) UpperCAmelCase = feature_extractor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(__lowerCamelCase , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = """This is a test string""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] , __lowerCamelCase : Optional[Any]=(2, 1_0, 1_6) , __lowerCamelCase : Union[str, Any]=7_7 ) -> List[str]: """simple docstring""" np.random.seed(__lowerCamelCase ) return np.random.rand(*__lowerCamelCase ) def _lowercase ( self : int ) -> List[str]: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits(shape=(1_0, 1_6) , seed=1_3 ) UpperCAmelCase = processor.decode(__lowerCamelCase ) UpperCAmelCase = decoder.decode_beams(__lowerCamelCase )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def _lowercase ( self : List[Any] , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCAmelCase = processor.batch_decode(__lowerCamelCase ) else: with get_context(__lowerCamelCase ).Pool() as pool: UpperCAmelCase = processor.batch_decode(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase = list(__lowerCamelCase ) with get_context("""fork""" ).Pool() as p: UpperCAmelCase = decoder.decode_beams_batch(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(__lowerCamelCase , decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] , decoded_processor.text ) self.assertListEqual(__lowerCamelCase , decoded_processor.logit_score ) self.assertListEqual(__lowerCamelCase , decoded_processor.lm_score ) def _lowercase ( self : int ) -> int: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = 1_5 UpperCAmelCase = -20.0 UpperCAmelCase = -4.0 UpperCAmelCase = processor.batch_decode( __lowerCamelCase , beam_width=__lowerCamelCase , beam_prune_logp=__lowerCamelCase , token_min_logp=__lowerCamelCase , ) UpperCAmelCase = decoded_processor_out.text UpperCAmelCase = list(__lowerCamelCase ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase = decoder.decode_beams_batch( __lowerCamelCase , __lowerCamelCase , beam_width=__lowerCamelCase , beam_prune_logp=__lowerCamelCase , token_min_logp=__lowerCamelCase , ) UpperCAmelCase = [d[0][0] for d in decoded_decoder_out] UpperCAmelCase = [d[0][2] for d in decoded_decoder_out] UpperCAmelCase = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] , __lowerCamelCase ) self.assertTrue(np.array_equal(__lowerCamelCase , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , __lowerCamelCase , atol=1e-3 ) ) self.assertTrue(np.array_equal(__lowerCamelCase , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9_474] , __lowerCamelCase , atol=1e-3 ) ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = 2.0 UpperCAmelCase = 5.0 UpperCAmelCase = -20.0 UpperCAmelCase = True UpperCAmelCase = processor.batch_decode( __lowerCamelCase , alpha=__lowerCamelCase , beta=__lowerCamelCase , unk_score_offset=__lowerCamelCase , lm_score_boundary=__lowerCamelCase , ) UpperCAmelCase = decoded_processor_out.text UpperCAmelCase = list(__lowerCamelCase ) decoder.reset_params( alpha=__lowerCamelCase , beta=__lowerCamelCase , unk_score_offset=__lowerCamelCase , lm_score_boundary=__lowerCamelCase , ) with get_context("""fork""" ).Pool() as pool: UpperCAmelCase = decoder.decode_beams_batch( __lowerCamelCase , __lowerCamelCase , ) UpperCAmelCase = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] , __lowerCamelCase ) UpperCAmelCase = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained(__lowerCamelCase ) UpperCAmelCase = processor.decoder.model_container[processor.decoder._model_key] UpperCAmelCase = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = os.listdir(__lowerCamelCase ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Dict ) -> Any: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = floats_list((3, 1_0_0_0) ) UpperCAmelCase = processor_wavaveca(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor_auto(__lowerCamelCase , return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2 ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = processor_wavaveca.batch_decode(__lowerCamelCase ) UpperCAmelCase = processor_auto.batch_decode(__lowerCamelCase ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def _lowercase ( self : Optional[Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.get_feature_extractor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_decoder() UpperCAmelCase = WavaVecaProcessorWithLM(tokenizer=__lowerCamelCase , feature_extractor=__lowerCamelCase , decoder=__lowerCamelCase ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , ) @staticmethod def _lowercase ( __lowerCamelCase : str , __lowerCamelCase : Tuple ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = [d[key] for d in offsets] return retrieved_list def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = self._get_dummy_logits()[0] UpperCAmelCase = processor.decode(__lowerCamelCase , output_word_offsets=__lowerCamelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(__lowerCamelCase , __lowerCamelCase ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """end_offset""" ) , [1, 3, 5] ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCAmelCase = self._get_dummy_logits() UpperCAmelCase = processor.batch_decode(__lowerCamelCase , output_word_offsets=__lowerCamelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(__lowerCamelCase , __lowerCamelCase ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(__lowerCamelCase , """word""" ) ) for o in outputs["""word_offsets"""]] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """end_offset""" ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" import torch UpperCAmelCase = load_dataset("""common_voice""" , """en""" , split="""train""" , streaming=__lowerCamelCase ) UpperCAmelCase = ds.cast_column("""audio""" , datasets.Audio(sampling_rate=1_6_0_0_0 ) ) UpperCAmelCase = iter(__lowerCamelCase ) UpperCAmelCase = next(__lowerCamelCase ) UpperCAmelCase = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCAmelCase = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCAmelCase = processor(sample["""audio"""]["""array"""] , return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCAmelCase = model(__lowerCamelCase ).logits.cpu().numpy() UpperCAmelCase = processor.decode(logits[0] , output_word_offsets=__lowerCamelCase ) UpperCAmelCase = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCAmelCase = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCAmelCase = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(__lowerCamelCase , """word""" ) ) , __lowerCamelCase ) self.assertEqual(""" """.join(self.get_from_offsets(__lowerCamelCase , """word""" ) ) , output.text ) # output times UpperCAmelCase = torch.tensor(self.get_from_offsets(__lowerCamelCase , """start_time""" ) ) UpperCAmelCase = torch.tensor(self.get_from_offsets(__lowerCamelCase , """end_time""" ) ) # fmt: off UpperCAmelCase = torch.tensor([1.4_199, 1.6_599, 2.2_599, 3.0, 3.24, 3.5_999, 3.7_999, 4.0_999, 4.26, 4.94, 5.28, 5.6_599, 5.78, 5.94, 6.32, 6.5_399, 6.6_599] ) UpperCAmelCase = torch.tensor([1.5_399, 1.8_999, 2.9, 3.16, 3.5_399, 3.72, 4.0_199, 4.1_799, 4.76, 5.1_599, 5.5_599, 5.6_999, 5.86, 6.1_999, 6.38, 6.6_199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=0.01 ) ) self.assertTrue(torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=0.01 ) )
627
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __lowercase : def __init__( self : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[Any]=1_3 , __lowerCamelCase : Optional[Any]=3_0 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : str=True , __lowerCamelCase : str=True , __lowerCamelCase : Union[str, Any]=3_2 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : Union[str, Any]=3_7 , __lowerCamelCase : str="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : str=1_0 , __lowerCamelCase : Union[str, Any]=0.02 , __lowerCamelCase : List[str]=3 , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : Any=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = type_sequence_label_size UpperCAmelCase = initializer_range UpperCAmelCase = scope UpperCAmelCase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase = (image_size // patch_size) ** 2 UpperCAmelCase = num_patches + 2 def _lowercase ( self : Tuple ) -> int: """simple docstring""" UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> str: """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase ( self : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = TFDeiTModel(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Any , __lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = TFDeiTForMaskedImageModeling(config=__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForMaskedImageModeling(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase ( self : Union[str, Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict , __lowerCamelCase : int ) -> Any: """simple docstring""" UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = TFDeiTForImageClassification(__lowerCamelCase ) UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase ( self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __lowercase ( __snake_case , __snake_case , unittest.TestCase ): UpperCamelCase = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDeiTModel, '''image-classification''': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def _lowercase ( self : str ) -> str: """simple docstring""" UpperCAmelCase = TFDeiTModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self : Union[str, Any] ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _lowercase ( self : List[Any] ) -> Dict: """simple docstring""" pass def _lowercase ( self : str ) -> List[Any]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCamelCase , tf.keras.layers.Dense ) ) def _lowercase ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__lowerCamelCase ) UpperCAmelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowercase ( self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCamelCase ) def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any=False ) -> int: """simple docstring""" UpperCAmelCase = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase ( self : Optional[int] ) -> str: """simple docstring""" for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = TFDeiTModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) ->Tuple: UpperCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __lowercase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _lowercase ( self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase = TFDeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=__lowerCamelCase , return_tensors="""tf""" ) # forward pass UpperCAmelCase = model(**__lowerCamelCase ) # verify the logits UpperCAmelCase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCAmelCase = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
627
1
import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path __a = [ {"""dataset""": """wikipedia""", """config_name""": """20220301.de"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.en"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.fr"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.frr"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.it"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.simple"""}, {"""dataset""": """snli""", """config_name""": """plain_text"""}, {"""dataset""": """eli5""", """config_name""": """LFQA_reddit"""}, {"""dataset""": """wiki40b""", """config_name""": """en"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.nq.compressed"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.nq.no_index"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.multiset.no_index"""}, {"""dataset""": """natural_questions""", """config_name""": """default"""}, ] def _UpperCamelCase ( lowerCAmelCase_=True ) ->Optional[int]: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__snake_case ) ) class __lowercase ( __snake_case ): UpperCamelCase = None UpperCamelCase = None def _lowercase ( self : List[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : Dict ) -> Tuple: """simple docstring""" with TemporaryDirectory() as tmp_dir: UpperCAmelCase = dataset_module_factory(__lowerCamelCase , cache_dir=__lowerCamelCase ) UpperCAmelCase = import_main_class(dataset_module.module_path , dataset=__lowerCamelCase ) UpperCAmelCase = builder_cls( cache_dir=__lowerCamelCase , config_name=__lowerCamelCase , hash=dataset_module.hash , ) UpperCAmelCase = """/""".join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=__lowerCamelCase ).replace(os.sep , """/""" ), config.DATASET_INFO_FILENAME, ] ) UpperCAmelCase = cached_path(__lowerCamelCase , cache_dir=__lowerCamelCase ) self.assertTrue(os.path.exists(__lowerCamelCase ) ) @pytest.mark.integration def _UpperCamelCase ( lowerCAmelCase_ ) ->Dict: UpperCAmelCase = tmp_path_factory.mktemp("""test_hf_gcp""" ) / """test_wikipedia_simple""" UpperCAmelCase = dataset_module_factory("""wikipedia""" , cache_dir=lowerCAmelCase_ ) UpperCAmelCase = import_main_class(dataset_module.module_path ) UpperCAmelCase = builder_cls( cache_dir=lowerCAmelCase_ , config_name="""20220301.frr""" , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam UpperCAmelCase = None builder_instance.download_and_prepare() UpperCAmelCase = builder_instance.as_dataset() assert ds @pytest.mark.integration def _UpperCamelCase ( lowerCAmelCase_ ) ->Tuple: UpperCAmelCase = dataset_module_factory("""wikipedia""" , cache_dir=lowerCAmelCase_ ) UpperCAmelCase = import_main_class(dataset_module.module_path , dataset=lowerCAmelCase_ ) UpperCAmelCase = builder_cls( cache_dir=lowerCAmelCase_ , config_name="""20220301.frr""" , hash=dataset_module.hash , ) UpperCAmelCase = builder_instance.as_streaming_dataset() assert ds assert isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) assert "train" in ds assert isinstance(ds["""train"""] , lowerCAmelCase_ ) assert next(iter(ds["""train"""] ) )
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class __lowercase ( unittest.TestCase ): def _lowercase ( self : List[str] ) -> Optional[int]: """simple docstring""" super().tearDown() gc.collect() def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """stabilityai/stable-diffusion-2""" , revision="""bf16""" , dtype=jnp.bfloataa , ) UpperCAmelCase = """A painting of a squirrel eating a burger""" UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = sd_pipe.prepare_inputs(__lowerCamelCase ) UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = jax.random.split(__lowerCamelCase , jax.device_count() ) UpperCAmelCase = sd_pipe(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , num_inference_steps=2_5 , jit=__lowerCamelCase )[0] assert images.shape == (jax.device_count(), 1, 7_6_8, 7_6_8, 3) UpperCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase = images[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] UpperCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase = jnp.array([0.4_238, 0.4_414, 0.4_395, 0.4_453, 0.4_629, 0.4_590, 0.4_531, 0.45_508, 0.4_512] ) print(F"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def _lowercase ( self : int ) -> str: """simple docstring""" UpperCAmelCase = """stabilityai/stable-diffusion-2""" UpperCAmelCase , UpperCAmelCase = FlaxDPMSolverMultistepScheduler.from_pretrained(__lowerCamelCase , subfolder="""scheduler""" ) UpperCAmelCase , UpperCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( __lowerCamelCase , scheduler=__lowerCamelCase , revision="""bf16""" , dtype=jnp.bfloataa , ) UpperCAmelCase = scheduler_params UpperCAmelCase = """A painting of a squirrel eating a burger""" UpperCAmelCase = jax.device_count() UpperCAmelCase = num_samples * [prompt] UpperCAmelCase = sd_pipe.prepare_inputs(__lowerCamelCase ) UpperCAmelCase = replicate(__lowerCamelCase ) UpperCAmelCase = shard(__lowerCamelCase ) UpperCAmelCase = jax.random.PRNGKey(0 ) UpperCAmelCase = jax.random.split(__lowerCamelCase , jax.device_count() ) UpperCAmelCase = sd_pipe(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , num_inference_steps=2_5 , jit=__lowerCamelCase )[0] assert images.shape == (jax.device_count(), 1, 7_6_8, 7_6_8, 3) UpperCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) UpperCAmelCase = images[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] UpperCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) UpperCAmelCase = jnp.array([0.4_336, 0.42_969, 0.4_453, 0.4_199, 0.4_297, 0.4_531, 0.4_434, 0.4_434, 0.4_297] ) print(F"""output_slice: {output_slice}""" ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
627
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __a = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
1
from __future__ import annotations __a = list[tuple[int, int]] __a = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __a = ([-1, 0], [0, -1], [1, 0], [0, 1]) # up, left, down, right class __lowercase : def __init__( self : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : float , __lowerCamelCase : Node | None , ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = pos_x UpperCAmelCase = pos_y UpperCAmelCase = (pos_y, pos_x) UpperCAmelCase = goal_x UpperCAmelCase = goal_y UpperCAmelCase = g_cost UpperCAmelCase = parent UpperCAmelCase = self.calculate_heuristic() def _lowercase ( self : List[str] ) -> float: """simple docstring""" UpperCAmelCase = abs(self.pos_x - self.goal_x ) UpperCAmelCase = abs(self.pos_y - self.goal_y ) return dx + dy def __lt__( self : Union[str, Any] , __lowerCamelCase : Dict ) -> bool: """simple docstring""" return self.f_cost < other.f_cost class __lowercase : def __init__( self : str , __lowerCamelCase : tuple[int, int] , __lowerCamelCase : tuple[int, int] ) -> str: """simple docstring""" UpperCAmelCase = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , __lowerCamelCase ) UpperCAmelCase = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9_9_9_9 , __lowerCamelCase ) UpperCAmelCase = [self.start] UpperCAmelCase = [] UpperCAmelCase = False def _lowercase ( self : List[Any] ) -> Path | None: """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: UpperCAmelCase = True return self.retrace_path(__lowerCamelCase ) self.closed_nodes.append(__lowerCamelCase ) UpperCAmelCase = self.get_successors(__lowerCamelCase ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(__lowerCamelCase ) else: # retrieve the best current path UpperCAmelCase = self.open_nodes.pop(self.open_nodes.index(__lowerCamelCase ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(__lowerCamelCase ) else: self.open_nodes.append(__lowerCamelCase ) if not self.reached: return [self.start.pos] return None def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Node ) -> list[Node]: """simple docstring""" UpperCAmelCase = [] for action in delta: UpperCAmelCase = parent.pos_x + action[1] UpperCAmelCase = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(__lowerCamelCase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( __lowerCamelCase , __lowerCamelCase , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , __lowerCamelCase , ) ) return successors def _lowercase ( self : Optional[Any] , __lowerCamelCase : Node | None ) -> Path: """simple docstring""" UpperCAmelCase = node UpperCAmelCase = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase = current_node.parent path.reverse() return path if __name__ == "__main__": __a = (0, 0) __a = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) print("""------""") __a = GreedyBestFirst(init, goal) __a = greedy_bf.search() if path: for pos_x, pos_y in path: __a = 2 for elem in grid: print(elem)
627
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging __a = logging.get_logger(__name__) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: try: with open(lowerCAmelCase_ , """rb""" ) as flax_state_f: UpperCAmelCase = from_bytes(lowerCAmelCase_ , flax_state_f.read() ) except UnpicklingError as e: try: with open(lowerCAmelCase_ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F"""Unable to convert {model_file} to Flax deserializable object. """ ) return load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Dict: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights UpperCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda lowerCAmelCase_ : x.dtype == jnp.bfloataa , lowerCAmelCase_ ) ).values() if any(lowerCAmelCase_ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) UpperCAmelCase = jax.tree_util.tree_map( lambda lowerCAmelCase_ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , lowerCAmelCase_ ) UpperCAmelCase = """""" UpperCAmelCase = flatten_dict(lowerCAmelCase_ , sep=""".""" ) UpperCAmelCase = pt_model.state_dict() # keep track of unexpected & missing keys UpperCAmelCase = [] UpperCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = jnp.transpose(lowerCAmelCase_ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] UpperCAmelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": UpperCAmelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(lowerCAmelCase_ ): UpperCAmelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) UpperCAmelCase = """.""".join(lowerCAmelCase_ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F"""Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected """ F"""to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) else: # add weight to pytorch dict UpperCAmelCase = np.asarray(lowerCAmelCase_ ) if not isinstance(lowerCAmelCase_ , np.ndarray ) else flax_tensor UpperCAmelCase = torch.from_numpy(lowerCAmelCase_ ) # remove from missing keys missing_keys.remove(lowerCAmelCase_ ) else: # weight is not expected by PyTorch model unexpected_keys.append(lowerCAmelCase_ ) pt_model.load_state_dict(lowerCAmelCase_ ) # re-transform missing_keys to list UpperCAmelCase = list(lowerCAmelCase_ ) if len(lowerCAmelCase_ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F""" {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing""" F""" {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture""" """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F""" IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect""" """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(lowerCAmelCase_ ) > 0: logger.warning( F"""Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly""" F""" initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to""" """ use it for predictions and inference.""" ) return pt_model
627
1
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
from __future__ import annotations from collections.abc import Sequence from typing import Literal def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->str | Literal[False]: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 UpperCAmelCase = """_""" if count > 1: return False else: return "".join(lowerCAmelCase_ ) def _UpperCamelCase ( lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] while True: UpperCAmelCase = ["""$"""] * len(lowerCAmelCase_ ) UpperCAmelCase = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): UpperCAmelCase = compare_string(binary[i] , binary[j] ) if k is False: UpperCAmelCase = """*""" UpperCAmelCase = """*""" temp.append("""X""" ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi UpperCAmelCase = list(set(lowerCAmelCase_ ) ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] for minterm in minterms: UpperCAmelCase = """""" for _ in range(lowerCAmelCase_ ): UpperCAmelCase = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) ->bool: UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = list(lowerCAmelCase_ ) UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[str]: UpperCAmelCase = [] UpperCAmelCase = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): UpperCAmelCase = 0 UpperCAmelCase = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 UpperCAmelCase = j if count == 1: UpperCAmelCase = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 temp.append(prime_implicants[i] ) while True: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = 0 for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = chart[i].count(1 ) if count_n > max_n: UpperCAmelCase = count_n UpperCAmelCase = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = 0 def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->list[list[int]]: UpperCAmelCase = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): UpperCAmelCase = prime_implicants[i].count("""_""" ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): UpperCAmelCase = 1 return chart def _UpperCamelCase ( ) ->None: UpperCAmelCase = int(input("""Enter the no. of variables\n""" ) ) UpperCAmelCase = [ float(lowerCAmelCase_ ) for x in input( """Enter the decimal representation of Minterms 'Spaces Separated'\n""" ).split() ] UpperCAmelCase = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = check(lowerCAmelCase_ ) print("""Prime Implicants are:""" ) print(lowerCAmelCase_ ) UpperCAmelCase = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print("""Essential Prime Implicants are:""" ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
627
1
from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
627
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class __lowercase ( unittest.TestCase ): UpperCamelCase = ViTImageProcessor if is_vision_available() else None @property def _lowercase ( self : Tuple ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = (3, 3_2, 1_2_8) UpperCAmelCase = tempfile.mkdtemp() # fmt: off UpperCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on UpperCAmelCase = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) UpperCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 3_2, """width""": 1_2_8}, } UpperCAmelCase = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] , **__lowerCamelCase : Union[str, Any] ) -> int: """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : int , **__lowerCamelCase : Optional[int] ) -> Optional[int]: """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Any ) -> str: """simple docstring""" UpperCAmelCase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) UpperCAmelCase = Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) return image_input def _lowercase ( self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_image_processor() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase = self.get_image_processor(do_normalize=__lowerCamelCase , padding_value=1.0 ) UpperCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _lowercase ( self : int ) -> Any: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCAmelCase = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _lowercase ( self : str ) -> int: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = processor(text=__lowerCamelCase ) UpperCAmelCase = tokenizer(__lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = """test""" UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """labels"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _lowercase ( self : str ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase = processor.char_decode(__lowerCamelCase ) UpperCAmelCase = tokenizer.batch_decode(__lowerCamelCase ) UpperCAmelCase = [seq.replace(""" """ , """""" ) for seq in decoded_tok] self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def _lowercase ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = None UpperCAmelCase = self.prepare_image_inputs() UpperCAmelCase = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = self.get_image_processor() UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = MgpstrProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_8 ) UpperCAmelCase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) UpperCAmelCase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) UpperCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""] )
627
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __a = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
627
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __a = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) __a = """sshleifer/student_marian_en_ro_6_1""" __a = """sshleifer/tiny-mbart""" @require_torch class __lowercase ( __snake_case ): def _lowercase ( self : Dict , __lowerCamelCase : List[Any]=False , __lowerCamelCase : str=None , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : str=True , __lowerCamelCase : List[str]=True , ) -> List[Any]: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=__lowerCamelCase , num_train_epochs=1 , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , predict_with_generate=__lowerCamelCase , do_train=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , ) UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history if not do_eval: return UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats UpperCAmelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowercase ( self : Dict ) -> str: """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowercase ( self : Tuple ) -> Any: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @require_torch_multi_gpu def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=__lowerCamelCase ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowercase ( self : Optional[int] ) -> Dict: """simple docstring""" self.run_seqaseq_quick( distributed=__lowerCamelCase , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=__lowerCamelCase ) @require_apex @require_torch_gpu def _lowercase ( self : str ) -> Optional[Any]: """simple docstring""" self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowerCamelCase , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowercase ( self : Union[str, Any] , __lowerCamelCase : Optional[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } UpperCAmelCase = experiments[experiment_id] UpperCAmelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} UpperCAmelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowerCamelCase , extra_args_str=data["""extra_args_str"""] ) UpperCAmelCase = len(re.findall(__lowerCamelCase , cl.err ) ) self.assertEqual(__lowerCamelCase , data["""n_matches"""] ) @slow def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=__lowerCamelCase , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(os.path.join(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = [log for log in logs if """eval_loss""" in log.keys()] UpperCAmelCase = eval_metrics[0] UpperCAmelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , __lowerCamelCase ) # test if do_predict saves generations and metrics UpperCAmelCase = os.listdir(__lowerCamelCase ) UpperCAmelCase = {os.path.basename(__lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowercase ( self : str ) -> int: """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowerCamelCase : str ) -> Tuple[int, float]: UpperCAmelCase = """--skip_memory_metrics 0""" UpperCAmelCase = self.run_trainer( max_len=1_2_8 , model_name=__lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=__lowerCamelCase , distributed=__lowerCamelCase , extra_args_str=__lowerCamelCase , do_eval=__lowerCamelCase , do_predict=__lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics UpperCAmelCase = TrainerState.load_from_json(Path(__lowerCamelCase , """trainer_state.json""" ) ).log_history UpperCAmelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) UpperCAmelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) UpperCAmelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) UpperCAmelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb UpperCAmelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig UpperCAmelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb UpperCAmelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings UpperCAmelCase = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" F""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowerCamelCase , __lowerCamelCase , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" F""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowerCamelCase , __lowerCamelCase , F"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _lowercase ( self : Any , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : float = 3e-3 , __lowerCamelCase : str = "adafactor" , __lowerCamelCase : bool = False , __lowerCamelCase : str = None , __lowerCamelCase : int = 0 , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : bool = True , __lowerCamelCase : int = None , ) -> Dict: """simple docstring""" UpperCAmelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" UpperCAmelCase = self.get_auto_remove_tmp_dir() UpperCAmelCase = F""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() UpperCAmelCase = F""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowerCamelCase )} """.split() UpperCAmelCase = """ --do_predict """.split() UpperCAmelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: UpperCAmelCase = get_gpu_count() UpperCAmelCase = get_torch_dist_unique_port() UpperCAmelCase = F""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() UpperCAmelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowerCamelCase , env=self.get_env() ) else: UpperCAmelCase = ["""run_translation.py"""] + args with patch.object(__lowerCamelCase , """argv""" , __lowerCamelCase ): main() return output_dir
627
1
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger __a = get_logger(__name__) class __lowercase ( enum.Enum ): UpperCamelCase = '''all_checks''' UpperCamelCase = '''basic_checks''' UpperCamelCase = '''no_checks''' class __lowercase ( __snake_case ): pass class __lowercase ( __snake_case ): pass class __lowercase ( __snake_case ): pass class __lowercase ( __snake_case ): pass def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ) ->Optional[Any]: if expected_checksums is None: logger.info("""Unable to verify checksums.""" ) return if len(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) ) if len(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) > 0: raise UnexpectedDownloadedFile(str(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) ) UpperCAmelCase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] UpperCAmelCase = """ for """ + verification_name if verification_name is not None else """""" if len(lowerCAmelCase_ ) > 0: raise NonMatchingChecksumError( F"""Checksums didn't match{for_verification_name}:\n""" F"""{bad_urls}\n""" """Set `verification_mode='no_checks'` to skip checksums verification and ignore this error""" ) logger.info("""All the checksums matched successfully""" + for_verification_name ) class __lowercase ( __snake_case ): pass class __lowercase ( __snake_case ): pass class __lowercase ( __snake_case ): pass class __lowercase ( __snake_case ): pass def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) ->Tuple: if expected_splits is None: logger.info("""Unable to verify splits sizes.""" ) return if len(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) > 0: raise ExpectedMoreSplits(str(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) ) if len(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) > 0: raise UnexpectedSplits(str(set(lowerCAmelCase_ ) - set(lowerCAmelCase_ ) ) ) UpperCAmelCase = [ {"""expected""": expected_splits[name], """recorded""": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(lowerCAmelCase_ ) > 0: raise NonMatchingSplitsSizesError(str(lowerCAmelCase_ ) ) logger.info("""All the splits matched successfully.""" ) def _UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = True ) ->dict: if record_checksum: UpperCAmelCase = shaaaa() with open(lowerCAmelCase_ , """rb""" ) as f: for chunk in iter(lambda: f.read(1 << 2_0 ) , b"""""" ): m.update(lowerCAmelCase_ ) UpperCAmelCase = m.hexdigest() else: UpperCAmelCase = None return {"num_bytes": os.path.getsize(lowerCAmelCase_ ), "checksum": checksum} def _UpperCamelCase ( lowerCAmelCase_ ) ->Tuple: if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
627
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class __lowercase ( unittest.TestCase ): def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = torch.nn.Linear(1_0 , 1_0 ) UpperCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase = Accelerator() UpperCAmelCase = accelerator.prepare(__lowerCamelCase ) try: pickle.loads(pickle.dumps(__lowerCamelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
627
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging __a = logging.get_logger(__name__) if is_vision_available(): import PIL class __lowercase ( __snake_case ): UpperCamelCase = ['''pixel_values'''] def __init__( self : List[str] , __lowerCamelCase : bool = True , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : PILImageResampling = PILImageResampling.BICUBIC , __lowerCamelCase : bool = True , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : bool = True , __lowerCamelCase : Union[int, float] = 1 / 2_5_5 , __lowerCamelCase : bool = True , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : bool = True , **__lowerCamelCase : Optional[Any] , ) -> None: """simple docstring""" super().__init__(**__lowerCamelCase ) UpperCAmelCase = size if size is not None else {"""shortest_edge""": 2_2_4} UpperCAmelCase = get_size_dict(__lowerCamelCase , default_to_square=__lowerCamelCase ) UpperCAmelCase = crop_size if crop_size is not None else {"""height""": 2_2_4, """width""": 2_2_4} UpperCAmelCase = get_size_dict(__lowerCamelCase , default_to_square=__lowerCamelCase , param_name="""crop_size""" ) UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = resample UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = do_normalize UpperCAmelCase = image_mean if image_mean is not None else OPENAI_CLIP_MEAN UpperCAmelCase = image_std if image_std is not None else OPENAI_CLIP_STD UpperCAmelCase = do_convert_rgb def _lowercase ( self : Dict , __lowerCamelCase : np.ndarray , __lowerCamelCase : Dict[str, int] , __lowerCamelCase : PILImageResampling = PILImageResampling.BICUBIC , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Tuple , ) -> np.ndarray: """simple docstring""" UpperCAmelCase = get_size_dict(__lowerCamelCase , default_to_square=__lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCAmelCase = get_resize_output_image_size(__lowerCamelCase , size=size["""shortest_edge"""] , default_to_square=__lowerCamelCase ) return resize(__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Any , __lowerCamelCase : np.ndarray , __lowerCamelCase : Dict[str, int] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Optional[int] , ) -> np.ndarray: """simple docstring""" UpperCAmelCase = get_size_dict(__lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(__lowerCamelCase , size=(size["""height"""], size["""width"""]) , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Optional[int] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Union[int, float] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Dict , ) -> List[Any]: """simple docstring""" return rescale(__lowerCamelCase , scale=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : Optional[Any] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Union[float, List[float]] , __lowerCamelCase : Union[float, List[float]] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Dict , ) -> np.ndarray: """simple docstring""" return normalize(__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def _lowercase ( self : str , __lowerCamelCase : ImageInput , __lowerCamelCase : bool = None , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : PILImageResampling = None , __lowerCamelCase : bool = None , __lowerCamelCase : int = None , __lowerCamelCase : bool = None , __lowerCamelCase : float = None , __lowerCamelCase : bool = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : Optional[Union[float, List[float]]] = None , __lowerCamelCase : bool = None , __lowerCamelCase : Optional[Union[str, TensorType]] = None , __lowerCamelCase : Optional[ChannelDimension] = ChannelDimension.FIRST , **__lowerCamelCase : Dict , ) -> PIL.Image.Image: """simple docstring""" UpperCAmelCase = do_resize if do_resize is not None else self.do_resize UpperCAmelCase = size if size is not None else self.size UpperCAmelCase = get_size_dict(__lowerCamelCase , param_name="""size""" , default_to_square=__lowerCamelCase ) UpperCAmelCase = resample if resample is not None else self.resample UpperCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase = crop_size if crop_size is not None else self.crop_size UpperCAmelCase = get_size_dict(__lowerCamelCase , param_name="""crop_size""" , default_to_square=__lowerCamelCase ) UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase = image_mean if image_mean is not None else self.image_mean UpperCAmelCase = image_std if image_std is not None else self.image_std UpperCAmelCase = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb UpperCAmelCase = make_list_of_images(__lowerCamelCase ) if not valid_images(__lowerCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: UpperCAmelCase = [convert_to_rgb(__lowerCamelCase ) for image in images] # All transformations expect numpy arrays. UpperCAmelCase = [to_numpy_array(__lowerCamelCase ) for image in images] if do_resize: UpperCAmelCase = [self.resize(image=__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase ) for image in images] if do_center_crop: UpperCAmelCase = [self.center_crop(image=__lowerCamelCase , size=__lowerCamelCase ) for image in images] if do_rescale: UpperCAmelCase = [self.rescale(image=__lowerCamelCase , scale=__lowerCamelCase ) for image in images] if do_normalize: UpperCAmelCase = [self.normalize(image=__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase ) for image in images] UpperCAmelCase = [to_channel_dimension_format(__lowerCamelCase , __lowerCamelCase ) for image in images] UpperCAmelCase = {"""pixel_values""": images} return BatchFeature(data=__lowerCamelCase , tensor_type=__lowerCamelCase )
627
from math import isqrt def _UpperCamelCase ( lowerCAmelCase_ ) ->bool: return all(number % divisor != 0 for divisor in range(2 , isqrt(lowerCAmelCase_ ) + 1 ) ) def _UpperCamelCase ( lowerCAmelCase_ = 1_0**6 ) ->int: UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 7 while prime_candidate < max_prime: primes_count += is_prime(lowerCAmelCase_ ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F"""{solution() = }""")
627
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __lowercase ( unittest.TestCase ): @slow def _lowercase ( self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) UpperCAmelCase = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" UpperCAmelCase = model(__lowerCamelCase )["""last_hidden_state"""] UpperCAmelCase = tf.TensorShape((1, 1_0, 7_6_8) ) self.assertEqual(output.shape , __lowerCamelCase ) # compare the actual values for a slice. UpperCAmelCase = tf.convert_to_tensor( [[[-0.0_254, 0.0_235, 0.1_027], [0.0_606, -0.1_811, -0.0_418], [-0.1_561, -0.1_127, 0.2_687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
627
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class __lowercase ( __snake_case ): UpperCamelCase = '''nllb-moe''' UpperCamelCase = ['''past_key_values'''] UpperCamelCase = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=1_2_8_1_1_2 , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : Optional[int]=1_2 , __lowerCamelCase : Union[str, Any]=4_0_9_6 , __lowerCamelCase : List[str]=1_6 , __lowerCamelCase : List[str]=1_2 , __lowerCamelCase : int=4_0_9_6 , __lowerCamelCase : Tuple=1_6 , __lowerCamelCase : str=0.05 , __lowerCamelCase : List[str]=0.05 , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : str="relu" , __lowerCamelCase : Dict=1_0_2_4 , __lowerCamelCase : List[str]=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : List[Any]=0.0 , __lowerCamelCase : Optional[Any]=0.02 , __lowerCamelCase : Dict=2 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=False , __lowerCamelCase : Tuple="float32" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=1_2_8 , __lowerCamelCase : List[str]=6_4 , __lowerCamelCase : List[Any]=4 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : str=0.001 , __lowerCamelCase : Optional[int]=0.001 , __lowerCamelCase : Tuple="all" , __lowerCamelCase : Any=False , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : List[str]=1.0 , __lowerCamelCase : Dict=0.2 , __lowerCamelCase : Union[str, Any]=1 , __lowerCamelCase : int=0 , __lowerCamelCase : Dict=2 , __lowerCamelCase : int=False , **__lowerCamelCase : str , ) -> int: """simple docstring""" UpperCAmelCase = vocab_size UpperCAmelCase = max_position_embeddings UpperCAmelCase = d_model UpperCAmelCase = encoder_ffn_dim UpperCAmelCase = encoder_layers UpperCAmelCase = encoder_attention_heads UpperCAmelCase = decoder_ffn_dim UpperCAmelCase = decoder_layers UpperCAmelCase = decoder_attention_heads UpperCAmelCase = dropout UpperCAmelCase = attention_dropout UpperCAmelCase = activation_dropout UpperCAmelCase = activation_function UpperCAmelCase = init_std UpperCAmelCase = encoder_layerdrop UpperCAmelCase = decoder_layerdrop UpperCAmelCase = use_cache UpperCAmelCase = encoder_layers UpperCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase = router_z_loss_coef UpperCAmelCase = router_aux_loss_coef UpperCAmelCase = decoder_sparse_step UpperCAmelCase = encoder_sparse_step UpperCAmelCase = num_experts UpperCAmelCase = expert_capacity UpperCAmelCase = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) UpperCAmelCase = router_dtype UpperCAmelCase = router_ignore_padding_tokens UpperCAmelCase = batch_prioritized_routing UpperCAmelCase = second_expert_policy UpperCAmelCase = normalize_router_prob_before_dropping UpperCAmelCase = moe_eval_capacity_token_fraction UpperCAmelCase = moe_token_dropout UpperCAmelCase = output_router_logits super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
627
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __a = {"""configuration_mra""": ["""MRA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MraConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ """MRA_PRETRAINED_MODEL_ARCHIVE_LIST""", """MraForMaskedLM""", """MraForMultipleChoice""", """MraForQuestionAnswering""", """MraForSequenceClassification""", """MraForTokenClassification""", """MraLayer""", """MraModel""", """MraPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
627
__a = [ (1000, """M"""), (900, """CM"""), (500, """D"""), (400, """CD"""), (100, """C"""), (90, """XC"""), (50, """L"""), (40, """XL"""), (10, """X"""), (9, """IX"""), (5, """V"""), (4, """IV"""), (1, """I"""), ] def _UpperCamelCase ( lowerCAmelCase_ ) ->int: UpperCAmelCase = {"""I""": 1, """V""": 5, """X""": 1_0, """L""": 5_0, """C""": 1_0_0, """D""": 5_0_0, """M""": 1_0_0_0} UpperCAmelCase = 0 UpperCAmelCase = 0 while place < len(lowerCAmelCase_ ): if (place + 1 < len(lowerCAmelCase_ )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def _UpperCamelCase ( lowerCAmelCase_ ) ->str: UpperCAmelCase = [] for arabic, roman in ROMAN: ((UpperCAmelCase) , (UpperCAmelCase)) = divmod(lowerCAmelCase_ , lowerCAmelCase_ ) result.append(roman * factor ) if number == 0: break return "".join(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
627
1