code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() lowercase : List[str] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) lowercase : List[str] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.encoder.norm.weight", "encoder.layernorm.weight"), ("transformer.encoder.norm.bias", "encoder.layernorm.bias"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ] ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> str: _snake_case = state_dict.pop(__A ) _snake_case = val def SCREAMING_SNAKE_CASE__ ( __A ) -> Any: _snake_case = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _snake_case = key.replace('backbone.0.body' , 'backbone.conv_encoder.model' ) _snake_case = value else: _snake_case = value return new_state_dict def SCREAMING_SNAKE_CASE__ ( __A ) -> List[str]: _snake_case = '' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _snake_case = state_dict.pop(F'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case = state_dict.pop(F'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case = in_proj_weight[:256, :] _snake_case = in_proj_bias[:256] _snake_case = in_proj_weight[256:512, :] _snake_case = in_proj_bias[256:512] _snake_case = in_proj_weight[-256:, :] _snake_case = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _snake_case = state_dict.pop(F'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case = state_dict.pop(F'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case = in_proj_weight[:256, :] _snake_case = in_proj_bias[:256] _snake_case = in_proj_weight[256:512, :] _snake_case = in_proj_bias[256:512] _snake_case = in_proj_weight[-256:, :] _snake_case = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _snake_case = state_dict.pop( F'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) _snake_case = state_dict.pop(F'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _snake_case = in_proj_weight_cross_attn[:256, :] _snake_case = in_proj_bias_cross_attn[:256] _snake_case = in_proj_weight_cross_attn[256:512, :] _snake_case = in_proj_bias_cross_attn[256:512] _snake_case = in_proj_weight_cross_attn[-256:, :] _snake_case = in_proj_bias_cross_attn[-256:] def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> Union[str, Any]: _snake_case , _snake_case = image.size _snake_case = max(__A , __A ) _snake_case = 800 if 'detection' in checkpoint_url else 1_000 _snake_case = target_max_size / current_max_size _snake_case = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def SCREAMING_SNAKE_CASE__ ( __A ) -> str: _snake_case = F.to_tensor(__A ) _snake_case = F.normalize(__A , mean=[0.4_8_5, 0.4_5_6, 0.4_0_6] , std=[0.2_2_9, 0.2_2_4, 0.2_2_5] ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Any: logger.info('Converting model...' ) # load original state dict _snake_case = torch.hub.load_state_dict_from_url(__A , map_location='cpu' ) # rename keys for src, dest in rename_keys: rename_key(__A , __A , __A ) _snake_case = rename_backbone_keys(__A ) # query, key and value matrices need special treatment read_in_q_k_v(__A ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _snake_case = 'model.' for key in state_dict.copy().keys(): if not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ): _snake_case = state_dict.pop(__A ) _snake_case = val # create HuggingFace model and load state dict _snake_case = TableTransformerConfig( backbone='resnet18' , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: _snake_case = 15 _snake_case = 2 _snake_case = {0: 'table', 1: 'table rotated'} _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} else: _snake_case = 125 _snake_case = 6 _snake_case = { 0: 'table', 1: 'table column', 2: 'table row', 3: 'table column header', 4: 'table projected row header', 5: 'table spanning cell', } _snake_case = idalabel _snake_case = {v: k for k, v in idalabel.items()} _snake_case = DetrImageProcessor( format='coco_detection' , max_size=800 if 'detection' in checkpoint_url else 1_000 ) _snake_case = TableTransformerForObjectDetection(__A ) model.load_state_dict(__A ) model.eval() # verify our conversion _snake_case = 'example_pdf.png' if 'detection' in checkpoint_url else 'example_table.png' _snake_case = hf_hub_download(repo_id='nielsr/example-pdf' , repo_type='dataset' , filename=__A ) _snake_case = Image.open(__A ).convert('RGB' ) _snake_case = normalize(resize(__A , __A ) ).unsqueeze(0 ) _snake_case = model(__A ) if "detection" in checkpoint_url: _snake_case = (1, 15, 3) _snake_case = torch.tensor( [[-6.7_8_9_7, -1_6.9_9_8_5, 6.7_9_3_7], [-8.0_1_8_6, -2_2.2_1_9_2, 6.9_6_7_7], [-7.3_1_1_7, -2_1.0_7_0_8, 7.4_0_5_5]] ) _snake_case = torch.tensor([[0.4_8_6_7, 0.1_7_6_7, 0.6_7_3_2], [0.6_7_1_8, 0.4_4_7_9, 0.3_8_3_0], [0.4_7_1_6, 0.1_7_6_0, 0.6_3_6_4]] ) else: _snake_case = (1, 125, 7) _snake_case = torch.tensor( [[-1_8.1_4_3_0, -8.3_2_1_4, 4.8_2_7_4], [-1_8.4_6_8_5, -7.1_3_6_1, -4.2_6_6_7], [-2_6.3_6_9_3, -9.3_4_2_9, -4.9_9_6_2]] ) _snake_case = torch.tensor([[0.4_9_8_3, 0.5_5_9_5, 0.9_4_4_0], [0.4_9_1_6, 0.6_3_1_5, 0.5_9_5_4], [0.6_1_0_8, 0.8_6_3_7, 0.1_1_3_5]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , __A , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , __A , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(__A ).mkdir(exist_ok=__A ) model.save_pretrained(__A ) image_processor.save_pretrained(__A ) if push_to_hub: # Push model to HF hub logger.info('Pushing model to the hub...' ) _snake_case = ( 'microsoft/table-transformer-detection' if 'detection' in checkpoint_url else 'microsoft/table-transformer-structure-recognition' ) model.push_to_hub(__A ) image_processor.push_to_hub(__A ) if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument( "--checkpoint_url", default="https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth", type=str, choices=[ "https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth", "https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth", ], help="URL of the Table Transformer checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) lowercase : List[Any] = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
42
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = {"configuration_mbart": ["MBART_PRETRAINED_CONFIG_ARCHIVE_MAP", "MBartConfig", "MBartOnnxConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ["MBartTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ["MBartTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ "MBART_PRETRAINED_MODEL_ARCHIVE_LIST", "MBartForCausalLM", "MBartForConditionalGeneration", "MBartForQuestionAnswering", "MBartForSequenceClassification", "MBartModel", "MBartPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ "TFMBartForConditionalGeneration", "TFMBartModel", "TFMBartPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ "FlaxMBartForConditionalGeneration", "FlaxMBartForQuestionAnswering", "FlaxMBartForSequenceClassification", "FlaxMBartModel", "FlaxMBartPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
158
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class A_ : '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = XGLMConfig SCREAMING_SNAKE_CASE__ : List[str] = {} SCREAMING_SNAKE_CASE__ : str = """gelu""" def __init__( self , lowercase_ , lowercase_=14 , lowercase_=7 , lowercase_=True , lowercase_=True , lowercase_=True , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=512 , lowercase_=0.02 , ): """simple docstring""" UpperCAmelCase_ : List[Any] = parent UpperCAmelCase_ : Optional[Any] = batch_size UpperCAmelCase_ : Dict = seq_length UpperCAmelCase_ : int = is_training UpperCAmelCase_ : Dict = use_input_mask UpperCAmelCase_ : Optional[int] = use_labels UpperCAmelCase_ : str = vocab_size UpperCAmelCase_ : int = d_model UpperCAmelCase_ : Any = num_hidden_layers UpperCAmelCase_ : Tuple = num_attention_heads UpperCAmelCase_ : Optional[Any] = ffn_dim UpperCAmelCase_ : List[str] = activation_function UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : List[str] = attention_dropout UpperCAmelCase_ : int = max_position_embeddings UpperCAmelCase_ : Optional[int] = initializer_range UpperCAmelCase_ : int = None UpperCAmelCase_ : Any = 0 UpperCAmelCase_ : Union[str, Any] = 2 UpperCAmelCase_ : Optional[int] = 1 def UpperCamelCase__ ( self ): """simple docstring""" return XGLMConfig.from_pretrained("facebook/xglm-564M" ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) UpperCAmelCase_ : Tuple = None if self.use_input_mask: UpperCAmelCase_ : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ : str = self.get_config() UpperCAmelCase_ : Union[str, Any] = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def UpperCamelCase__ ( self ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=lowercase_ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = self.prepare_config_and_inputs() ( UpperCAmelCase_ ) : List[str] = config_and_inputs UpperCAmelCase_ : str = { """input_ids""": input_ids, """head_mask""": head_mask, } return config, inputs_dict @require_tf class A_ (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : List[Any] = (TFXGLMForCausalLM,) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : Tuple = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE__ : List[str] = False SCREAMING_SNAKE_CASE__ : Optional[int] = False SCREAMING_SNAKE_CASE__ : Union[str, Any] = False def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = TFXGLMModelTester(self ) UpperCAmelCase_ : List[str] = ConfigTester(self , config_class=lowercase_ , n_embd=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Optional[int] = TFXGLMModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) @unittest.skip(reason="Currently, model embeddings are going to undergo a major refactor." ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class A_ (unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self , lowercase_=True ): """simple docstring""" UpperCAmelCase_ : Optional[int] = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : Optional[Any] = tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off UpperCAmelCase_ : Optional[Any] = [2, 268, 9865, 67, 11, 1988, 5_7252, 9865, 5, 984, 67, 1988, 21_3838, 1658, 53, 7_0446, 33, 6657, 278, 1581] # fmt: on UpperCAmelCase_ : List[str] = model.generate(lowercase_ , do_sample=lowercase_ , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , lowercase_ ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : Any = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) tf.random.set_seed(0 ) UpperCAmelCase_ : Union[str, Any] = tokenizer("Today is a nice day and" , return_tensors="tf" ) UpperCAmelCase_ : Optional[Any] = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(":/CPU:0" ): UpperCAmelCase_ : Any = model.generate(lowercase_ , do_sample=lowercase_ , seed=[7, 0] ) UpperCAmelCase_ : List[Any] = tokenizer.decode(output_ids[0] , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : Any = ( """Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due""" ) self.assertEqual(lowercase_ , lowercase_ ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : Tuple = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : int = """left""" # use different length sentences to test batching UpperCAmelCase_ : str = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When""", """Hello, my dog is a little""", ] UpperCAmelCase_ : Tuple = tokenizer(lowercase_ , return_tensors="tf" , padding=lowercase_ ) UpperCAmelCase_ : Dict = inputs["""input_ids"""] UpperCAmelCase_ : Any = model.generate(input_ids=lowercase_ , attention_mask=inputs["attention_mask"] , max_new_tokens=12 ) UpperCAmelCase_ : Optional[Any] = tokenizer(sentences[0] , return_tensors="tf" ).input_ids UpperCAmelCase_ : Union[str, Any] = model.generate(input_ids=lowercase_ , max_new_tokens=12 ) UpperCAmelCase_ : Dict = tokenizer(sentences[1] , return_tensors="tf" ).input_ids UpperCAmelCase_ : Optional[Any] = model.generate(input_ids=lowercase_ , max_new_tokens=12 ) UpperCAmelCase_ : Optional[Any] = tokenizer.batch_decode(lowercase_ , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : Optional[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : List[str] = tokenizer.decode(output_padded[0] , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : Dict = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When left padding is applied, the sequence will be """ """a single""", """Hello, my dog is a little bit of a shy one, but he is very friendly""", ] self.assertListEqual(lowercase_ , lowercase_ ) self.assertListEqual(lowercase_ , [non_padded_sentence, padded_sentence] )
371
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'vocab.json'} _a = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _a = {'mgp-str': 27} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowercase_ , lowercase_="[GO]" , lowercase_="[GO]" , lowercase_="[s]" , lowercase_="[GO]" , **lowercase_ ): """simple docstring""" super().__init__( unk_token=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , pad_token=lowercase_ , **lowercase_ , ) with open(lowercase_ , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ : Dict = json.load(lowercase_ ) UpperCAmelCase_ : Dict = {v: k for k, v in self.vocab.items()} @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.vocab ) def UpperCamelCase__ ( self ): """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [] for s in text: char_tokens.extend(lowercase_ ) return char_tokens def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.vocab.get(lowercase_ , self.vocab.get(self.unk_token ) ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.decoder.get(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(lowercase_ ): logger.error("Vocabulary path ({}) should be a directory".format(lowercase_ ) ) return UpperCAmelCase_ : Optional[int] = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowercase_ , ensure_ascii=lowercase_ ) + "\n" ) return (vocab_file,)
23
0
"""simple docstring""" import os import zipfile import pytest from datasets.utils.extract import ( BzipaExtractor, Extractor, GzipExtractor, LzaExtractor, SevenZipExtractor, TarExtractor, XzExtractor, ZipExtractor, ZstdExtractor, ) from .utils import require_lza, require_pyazr, require_zstandard @pytest.mark.parametrize( '''compression_format, is_archive''' , [ ('''7z''', True), ('''bz2''', False), ('''gzip''', False), ('''lz4''', False), ('''tar''', True), ('''xz''', False), ('''zip''', True), ('''zstd''', False), ] , ) def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): __lowercase : List[Any] = { '''7z''': (seven_zip_file, SevenZipExtractor), '''bz2''': (bza_file, BzipaExtractor), '''gzip''': (gz_file, GzipExtractor), '''lz4''': (lza_file, LzaExtractor), '''tar''': (tar_file, TarExtractor), '''xz''': (xz_file, XzExtractor), '''zip''': (zip_file, ZipExtractor), '''zstd''': (zstd_file, ZstdExtractor), } __lowercase ,__lowercase : Union[str, Any] = input_paths_and_base_extractors[compression_format] if input_path is None: __lowercase : Optional[Any] = f"""for '{compression_format}' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(__UpperCamelCase ) assert base_extractor.is_extractable(__UpperCamelCase ) __lowercase : Union[str, Any] = tmp_path / ('''extracted''' if is_archive else '''extracted.txt''') base_extractor.extract(__UpperCamelCase , __UpperCamelCase ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name __lowercase : Dict = file_path.read_text(encoding='''utf-8''' ) else: __lowercase : str = output_path.read_text(encoding='''utf-8''' ) __lowercase : Tuple = text_file.read_text(encoding='''utf-8''' ) assert extracted_file_content == expected_file_content @pytest.mark.parametrize( '''compression_format, is_archive''' , [ ('''7z''', True), ('''bz2''', False), ('''gzip''', False), ('''lz4''', False), ('''tar''', True), ('''xz''', False), ('''zip''', True), ('''zstd''', False), ] , ) def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): __lowercase : List[str] = { '''7z''': seven_zip_file, '''bz2''': bza_file, '''gzip''': gz_file, '''lz4''': lza_file, '''tar''': tar_file, '''xz''': xz_file, '''zip''': zip_file, '''zstd''': zstd_file, } __lowercase : Tuple = input_paths[compression_format] if input_path is None: __lowercase : Optional[Any] = f"""for '{compression_format}' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(__UpperCamelCase ) __lowercase : Optional[int] = Extractor.infer_extractor_format(__UpperCamelCase ) assert extractor_format is not None __lowercase : Optional[Any] = tmp_path / ('''extracted''' if is_archive else '''extracted.txt''') Extractor.extract(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name __lowercase : str = file_path.read_text(encoding='''utf-8''' ) else: __lowercase : Any = output_path.read_text(encoding='''utf-8''' ) __lowercase : Dict = text_file.read_text(encoding='''utf-8''' ) assert extracted_file_content == expected_file_content @pytest.fixture def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase ): import tarfile __lowercase : Tuple = tmp_path / '''data_dot_dot''' directory.mkdir() __lowercase : Tuple = directory / '''tar_file_with_dot_dot.tar''' with tarfile.TarFile(__UpperCamelCase , '''w''' ) as f: f.add(__UpperCamelCase , arcname=os.path.join('''..''' , text_file.name ) ) return path @pytest.fixture def __UpperCAmelCase ( __UpperCamelCase ): import tarfile __lowercase : Tuple = tmp_path / '''data_sym_link''' directory.mkdir() __lowercase : List[str] = directory / '''tar_file_with_sym_link.tar''' os.symlink('''..''' , directory / '''subdir''' , target_is_directory=__UpperCamelCase ) with tarfile.TarFile(__UpperCamelCase , '''w''' ) as f: f.add(str(directory / '''subdir''' ) , arcname='''subdir''' ) # str required by os.readlink on Windows and Python < 3.8 return path @pytest.mark.parametrize( '''insecure_tar_file, error_log''' , [('''tar_file_with_dot_dot''', '''illegal path'''), ('''tar_file_with_sym_link''', '''Symlink''')] , ) def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): __lowercase : List[str] = { '''tar_file_with_dot_dot''': tar_file_with_dot_dot, '''tar_file_with_sym_link''': tar_file_with_sym_link, } __lowercase : Union[str, Any] = insecure_tar_files[insecure_tar_file] __lowercase : Dict = tmp_path / '''extracted''' TarExtractor.extract(__UpperCamelCase , __UpperCamelCase ) assert caplog.text for record in caplog.records: assert record.levelname == "ERROR" assert error_log in record.msg def __UpperCAmelCase ( __UpperCamelCase ): # We should have less false positives than zipfile.is_zipfile # We do that by checking only the magic number __lowercase : Tuple = tmpdir / '''not_a_zip_file''' # From: https://github.com/python/cpython/pull/5053 __lowercase : Tuple = ( B'''\x89PNG\r\n\x1a\n\x00\x00\x00\rIHDR\x00\x00\x00\x01\x00\x00''' B'''\x00\x02\x08\x06\x00\x00\x00\x99\x81\xb6\'\x00\x00\x00\x15I''' B'''DATx\x01\x01\n\x00\xf5\xff\x00PK\x05\x06\x00PK\x06\x06\x07''' B'''\xac\x01N\xc6|a\r\x00\x00\x00\x00IEND\xaeB`\x82''' ) with not_a_zip_file.open('''wb''' ) as f: f.write(__UpperCamelCase ) assert zipfile.is_zipfile(str(__UpperCamelCase ) ) # is a false positive for `zipfile` assert not ZipExtractor.is_extractable(__UpperCamelCase ) # but we're right
249
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm a_ = logging.get_logger(__name__) @dataclass class UpperCAmelCase_ ( snake_case ): UpperCamelCase =[ "no_inference", "no_cuda", "no_tpu", "no_speed", "no_memory", "no_env_print", "no_multi_process", ] def __init__( self , **UpperCamelCase_ ) -> Optional[Any]: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __lowercase : Union[str, Any] = deprecated_arg[3:] setattr(self , UpperCamelCase_ , not kwargs.pop(UpperCamelCase_ ) ) logger.warning( F"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" F""" {positive_arg}={kwargs[positive_arg]}""" ) __lowercase : Dict = kwargs.pop('''torchscript''' , self.torchscript ) __lowercase : str = kwargs.pop('''torch_xla_tpu_print_metrics''' , self.torch_xla_tpu_print_metrics ) __lowercase : str = kwargs.pop('''fp16_opt_level''' , self.fpaa_opt_level ) super().__init__(**UpperCamelCase_ ) UpperCamelCase =field(default=snake_case , metadata={"help": "Trace the models using torchscript"} ) UpperCamelCase =field(default=snake_case , metadata={"help": "Print Xla/PyTorch tpu metrics"} ) UpperCamelCase =field( default="O1" , metadata={ "help": ( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']. " "See details at https://nvidia.github.io/apex/amp.html" ) } , ) @cached_property def _lowerCamelCase ( self ) -> Tuple["torch.device", int]: requires_backends(self , ['''torch'''] ) logger.info('''PyTorch: setting up devices''' ) if not self.cuda: __lowercase : str = torch.device('''cpu''' ) __lowercase : Optional[Any] = 0 elif is_torch_tpu_available(): __lowercase : str = xm.xla_device() __lowercase : Any = 0 else: __lowercase : List[str] = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) __lowercase : Any = torch.cuda.device_count() return device, n_gpu @property def _lowerCamelCase ( self ) -> Union[str, Any]: return is_torch_tpu_available() and self.tpu @property def _lowerCamelCase ( self ) -> int: requires_backends(self , ['''torch'''] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def _lowerCamelCase ( self ) -> "torch.device": requires_backends(self , ['''torch'''] ) return self._setup_devices[0] @property def _lowerCamelCase ( self ) -> int: requires_backends(self , ['''torch'''] ) return self._setup_devices[1] @property def _lowerCamelCase ( self ) -> Dict: return self.n_gpu > 0
249
1
from __future__ import annotations from random import random class UpperCamelCase : def __init__( self , UpperCAmelCase__ = None ): A__ = value A__ = random() A__ = None A__ = None def __repr__( self ): from pprint import pformat if self.left is None and self.right is None: return F"""'{self.value}: {self.prior:.5}'""" else: return pformat( {F"""{self.value}: {self.prior:.5}""": (self.left, self.right)} , indent=1 ) def __str__( self ): A__ = str(self.value ) + " " A__ = str(self.left or "" ) A__ = str(self.right or "" ) return value + left + right def UpperCamelCase ( _A : Node | None , _A : int )-> tuple[Node | None, Node | None]: """simple docstring""" if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: A__ , A__ = split(root.left , _A ) return left, root else: A__ , A__ = split(root.right , _A ) return root, right def UpperCamelCase ( _A : Node | None , _A : Node | None )-> Node | None: """simple docstring""" if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: A__ = merge(left.right , _A ) return left else: A__ = merge(_A , right.left ) return right def UpperCamelCase ( _A : Node | None , _A : int )-> Node | None: """simple docstring""" A__ = Node(_A ) A__ , A__ = split(_A , _A ) return merge(merge(_A , _A ) , _A ) def UpperCamelCase ( _A : Node | None , _A : int )-> Node | None: """simple docstring""" A__ , A__ = split(_A , value - 1 ) A__ , A__ = split(_A , _A ) return merge(_A , _A ) def UpperCamelCase ( _A : Node | None )-> None: """simple docstring""" if not root: # None return else: inorder(root.left ) print(root.value , end="," ) inorder(root.right ) def UpperCamelCase ( _A : Node | None , _A : str )-> Node | None: """simple docstring""" for arg in args.split(): if arg[0] == "+": A__ = insert(_A , int(arg[1:] ) ) elif arg[0] == "-": A__ = erase(_A , int(arg[1:] ) ) else: print("Unknown command" ) return root def UpperCamelCase ( )-> None: """simple docstring""" A__ = None print( "enter numbers to create a tree, + value to add value into treap, " "- value to erase all nodes with value. 'q' to quit. " ) A__ = input() while args != "q": A__ = interact_treap(_A , _A ) print(_A ) A__ = input() print("good by!" ) if __name__ == "__main__": import doctest doctest.testmod() main()
198
import datasets from .evaluate import evaluate UpperCAmelCase_ : List[Any] = "\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n" UpperCAmelCase_ : Any = "\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n" UpperCAmelCase_ : Tuple = "\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair as given in the references (see below)\n - 'prediction_text': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair (see above),\n - 'answers': a Dict in the SQuAD dataset format\n {\n 'text': list of possible texts for the answer, as a list of strings\n 'answer_start': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n 'exact_match': Exact match (the normalized answer exactly match the gold answer)\n 'f1': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{'prediction_text': '1976', 'id': '56e10a3be3433e1400422b22'}]\n >>> references = [{'answers': {'answer_start': [97], 'text': ['1976']}, 'id': '56e10a3be3433e1400422b22'}]\n >>> squad_metric = datasets.load_metric(\"squad\")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 100.0, 'f1': 100.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase ( datasets.Metric ): def __A ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": {"id": datasets.Value("string" ), "prediction_text": datasets.Value("string" )}, "references": { "id": datasets.Value("string" ), "answers": datasets.features.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), }, } ) , codebase_urls=["https://rajpurkar.github.io/SQuAD-explorer/"] , reference_urls=["https://rajpurkar.github.io/SQuAD-explorer/"] , ) def __A ( self , UpperCAmelCase__ , UpperCAmelCase__ ): A__ = {prediction["id"]: prediction["prediction_text"] for prediction in predictions} A__ = [ { "paragraphs": [ { "qas": [ { "answers": [{"text": answer_text} for answer_text in ref["answers"]["text"]], "id": ref["id"], } for ref in references ] } ] } ] A__ = evaluate(dataset=UpperCAmelCase__ , predictions=UpperCAmelCase__ ) return score
198
1
from ..utils import DummyObject, requires_backends class A__(metaclass=a_ ): """simple docstring""" _A : Optional[Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Tuple: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : List[Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Any: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : str = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> int: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Union[str, Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[int]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Optional[Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> int: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Optional[Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Union[str, Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Optional[int] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Union[str, Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : List[Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Dict: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : List[str] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> int: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : str = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[int]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : List[str] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> str: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : List[str] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> List[Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Any = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : str = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[int]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : int = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[int]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Tuple = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> str: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : int = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> str: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : int = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Any: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Union[str, Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> int: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : List[str] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : str = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Tuple: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Optional[Any] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Dict = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Union[str, Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Any = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Union[str, Any]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Tuple = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Any: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Optional[int] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Tuple: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : int = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[int]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : Optional[int] = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> int: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : int = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[int]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : int = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> List[str]: requires_backends(self , ["""sentencepiece"""] ) class A__(metaclass=a_ ): """simple docstring""" _A : str = ['''sentencepiece'''] def __init__( self , *_lowercase , **_lowercase ) -> Optional[Any]: requires_backends(self , ["""sentencepiece"""] )
248
__snake_case : Dict = 6_55_21 def _UpperCAmelCase ( a__): '''simple docstring''' a_ : List[str] = 1 a_ : Any = 0 for plain_chr in plain_text: a_ : Tuple = (a + ord(a__)) % MOD_ADLER a_ : Union[str, Any] = (b + a) % MOD_ADLER return (b << 1_6) | a
248
1
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex a__ = logging.getLogger(__name__) class UpperCAmelCase_ : """simple docstring""" def __init__( self ) -> Union[str, Any]: _a : str = False def __lowercase ( self , _a , _a , _a , _a ) -> int: if not self.initialized: _a : Optional[Any] = RagRetriever( _a , question_encoder_tokenizer=_a , generator_tokenizer=_a , index=_a , init_retrieval=_a , ) _a : Optional[int] = True def __lowercase ( self ) -> str: self.retriever.index.init_index() def __lowercase ( self , _a , _a ) -> Any: _a , _a : str = self.retriever._main_retrieve(_a , _a ) return doc_ids, retrieved_doc_embeds class UpperCAmelCase_ ( __lowercase ): """simple docstring""" def __init__( self , _a , _a , _a , _a , _a=None ) -> Optional[int]: if index is not None and index.is_initialized() and len(_a ) > 0: raise ValueError( '''When using Ray for distributed fine-tuning, ''' '''you\'ll need to provide the paths instead, ''' '''as the dataset and the index are loaded ''' '''separately. More info in examples/rag/use_own_knowledge_dataset.py ''' ) super().__init__( _a , question_encoder_tokenizer=_a , generator_tokenizer=_a , index=_a , init_retrieval=_a , ) _a : List[Any] = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(_a , _a , _a , _a ) for worker in self.retrieval_workers ] ) def __lowercase ( self ) -> str: logger.info('''initializing retrieval''' ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def __lowercase ( self , _a , _a ) -> int: if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. _a : Optional[int] = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] _a , _a : Optional[int] = ray.get(random_worker.retrieve.remote(_a , _a ) ) else: _a , _a : List[Any] = self._main_retrieve(_a , _a ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(_a ) @classmethod def __lowercase ( cls , _a , _a=None , **_a ) -> Optional[int]: return super(_a , cls ).get_tokenizers(_a , _a , **_a ) @classmethod def __lowercase ( cls , _a , _a , _a=None , **_a ) -> List[str]: _a : str = kwargs.pop('''config''' , _a ) or RagConfig.from_pretrained(_a , **_a ) _a : str = RagTokenizer.from_pretrained(_a , config=_a ) _a : int = rag_tokenizer.question_encoder _a : Union[str, Any] = rag_tokenizer.generator if indexed_dataset is not None: _a : List[Any] = '''custom''' _a : Any = CustomHFIndex(config.retrieval_vector_size , _a ) else: _a : str = cls._build_index(_a ) return cls( _a , question_encoder_tokenizer=_a , generator_tokenizer=_a , retrieval_workers=_a , index=_a , )
15
from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch a__ = logging.get_logger(__name__) @add_end_docstrings( __lowercase , r"\n top_k (`int`, defaults to 5):\n The number of predictions to return.\n targets (`str` or `List[str]`, *optional*):\n When passed, the model will limit the scores to the passed targets instead of looking up in the whole\n vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting\n token will be used (with a warning, and that might be slower).\n\n " , ) class UpperCAmelCase_ ( __lowercase ): """simple docstring""" def __lowercase ( self , _a ) -> np.ndarray: if self.framework == "tf": _a : List[str] = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": _a : Tuple = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=_a ) else: raise ValueError('''Unsupported framework''' ) return masked_index def __lowercase ( self , _a ) -> np.ndarray: _a : int = self.get_masked_index(_a ) _a : Tuple = np.prod(masked_index.shape ) if numel < 1: raise PipelineException( '''fill-mask''' , self.model.base_model_prefix , F"""No mask_token ({self.tokenizer.mask_token}) found on the input""" , ) def __lowercase ( self , _a ) -> Optional[int]: if isinstance(_a , _a ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input['''input_ids'''][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(_a ) def __lowercase ( self , _a , _a=None , **_a ) -> Dict[str, GenericTensor]: if return_tensors is None: _a : Union[str, Any] = self.framework _a : str = self.tokenizer(_a , return_tensors=_a ) self.ensure_exactly_one_mask_token(_a ) return model_inputs def __lowercase ( self , _a ) -> Optional[Any]: _a : List[str] = self.model(**_a ) _a : Any = model_inputs['''input_ids'''] return model_outputs def __lowercase ( self , _a , _a=5 , _a=None ) -> str: # Cap top_k if there are targets if target_ids is not None and target_ids.shape[0] < top_k: _a : List[Any] = target_ids.shape[0] _a : Any = model_outputs['''input_ids'''][0] _a : List[str] = model_outputs['''logits'''] if self.framework == "tf": _a : Tuple = tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] _a : List[str] = outputs.numpy() _a : Dict = outputs[0, masked_index, :] _a : str = stable_softmax(_a , axis=-1 ) if target_ids is not None: _a : Any = tf.gather_nd(tf.squeeze(_a , 0 ) , target_ids.reshape(-1 , 1 ) ) _a : Union[str, Any] = tf.expand_dims(_a , 0 ) _a : Optional[int] = tf.math.top_k(_a , k=_a ) _a , _a : Optional[Any] = topk.values.numpy(), topk.indices.numpy() else: _a : Optional[Any] = torch.nonzero(input_ids == self.tokenizer.mask_token_id , as_tuple=_a ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample _a : List[str] = outputs[0, masked_index, :] _a : List[Any] = logits.softmax(dim=-1 ) if target_ids is not None: _a : List[Any] = probs[..., target_ids] _a , _a : Optional[Any] = probs.topk(_a ) _a : Dict = [] _a : List[Any] = values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist() , predictions.tolist() ) ): _a : Optional[Any] = [] for v, p in zip(_values , _predictions ): # Copy is important since we're going to modify this array in place _a : Optional[int] = input_ids.numpy().copy() if target_ids is not None: _a : Tuple = target_ids[p].tolist() _a : List[str] = p # Filter padding out: _a : List[Any] = tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back _a : List[str] = self.tokenizer.decode(_a , skip_special_tokens=_a ) _a : List[Any] = {'''score''': v, '''token''': p, '''token_str''': self.tokenizer.decode([p] ), '''sequence''': sequence} row.append(_a ) result.append(_a ) if single_mask: return result[0] return result def __lowercase ( self , _a , _a=None ) -> Dict: if isinstance(_a , _a ): _a : Tuple = [targets] try: _a : int = self.tokenizer.get_vocab() except Exception: _a : Any = {} _a : List[Any] = [] for target in targets: _a : List[Any] = vocab.get(_a , _a ) if id_ is None: _a : Tuple = self.tokenizer( _a , add_special_tokens=_a , return_attention_mask=_a , return_token_type_ids=_a , max_length=1 , truncation=_a , )['''input_ids'''] if len(_a ) == 0: logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ '''We cannot replace it with anything meaningful, ignoring it''' ) continue _a : Tuple = input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( F"""The specified target token `{target}` does not exist in the model vocabulary. """ F"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) _a : List[str] = list(set(_a ) ) if len(_a ) == 0: raise ValueError('''At least one target must be provided when passed.''' ) _a : int = np.array(_a ) return target_ids def __lowercase ( self , _a=None , _a=None ) -> Tuple: _a : str = {} if targets is not None: _a : List[Any] = self.get_target_ids(_a , _a ) _a : Optional[Any] = target_ids if top_k is not None: _a : Union[str, Any] = top_k if self.tokenizer.mask_token_id is None: raise PipelineException( '''fill-mask''' , self.model.base_model_prefix , '''The tokenizer does not define a `mask_token`.''' ) return {}, {}, postprocess_params def __call__( self , _a , *_a , **_a ) -> int: _a : Optional[Any] = super().__call__(_a , **_a ) if isinstance(_a , _a ) and len(_a ) == 1: return outputs[0] return outputs
15
1
from __future__ import annotations from collections.abc import Iterator from typing import Any class lowercase__: """simple docstring""" def __init__( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Any ) -> Any: lowercase_ = data lowercase_ = None class lowercase__: """simple docstring""" def __init__( self : Any ) -> List[Any]: lowercase_ = None lowercase_ = None def __iter__( self : int ) -> Iterator[Any]: lowercase_ = self.head while self.head: yield node.data lowercase_ = node.next if node == self.head: break def __len__( self : str ) -> int: return sum(1 for _ in self ) def __repr__( self : Optional[Any] ) -> Tuple: return "->".join(str(SCREAMING_SNAKE_CASE_ ) for item in iter(self ) ) def _lowercase ( self : Optional[Any] , SCREAMING_SNAKE_CASE_ : Any ) -> None: self.insert_nth(len(self ) , SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : str , SCREAMING_SNAKE_CASE_ : Any ) -> None: self.insert_nth(0 , SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : List[str] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Any ) -> None: if index < 0 or index > len(self ): raise IndexError('''list index out of range.''' ) lowercase_ = Node(SCREAMING_SNAKE_CASE_ ) if self.head is None: lowercase_ = new_node # first node points itself lowercase_ = lowercase_ = new_node elif index == 0: # insert at head lowercase_ = self.head lowercase_ = lowercase_ = new_node else: lowercase_ = self.head for _ in range(index - 1 ): lowercase_ = temp.next lowercase_ = temp.next lowercase_ = new_node if index == len(self ) - 1: # insert at tail lowercase_ = new_node def _lowercase ( self : int ) -> List[Any]: return self.delete_nth(0 ) def _lowercase ( self : str ) -> Any: return self.delete_nth(len(self ) - 1 ) def _lowercase ( self : Tuple , SCREAMING_SNAKE_CASE_ : int = 0 ) -> Any: if not 0 <= index < len(self ): raise IndexError('''list index out of range.''' ) lowercase_ = self.head if self.head == self.tail: # just one node lowercase_ = lowercase_ = None elif index == 0: # delete head node lowercase_ = self.tail.next.next lowercase_ = self.head.next else: lowercase_ = self.head for _ in range(index - 1 ): lowercase_ = temp.next lowercase_ = temp.next lowercase_ = temp.next.next if index == len(self ) - 1: # delete at tail lowercase_ = temp return delete_node.data def _lowercase ( self : Union[str, Any] ) -> bool: return len(self ) == 0 def a ( ): '''simple docstring''' lowercase_ = CircularLinkedList() assert len(snake_case__ ) == 0 assert circular_linked_list.is_empty() is True assert str(snake_case__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(snake_case__ ) == i circular_linked_list.insert_nth(snake_case__ , i + 1 ) assert str(snake_case__ ) == "->".join(str(snake_case__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(snake_case__ ) == "->".join(str(snake_case__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(snake_case__ ) == "->".join(str(snake_case__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(snake_case__ ) == "->".join(str(snake_case__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(snake_case__ ) == "->".join(str(snake_case__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
30
'''simple docstring''' from maths.prime_check import is_prime def __lowerCAmelCase ( snake_case__ ): if not isinstance(snake_case__ , snake_case__ ): __UpperCamelCase : Optional[int] = F"Input value of [number={number}] must be an integer" raise TypeError(snake_case__ ) if is_prime(snake_case__ ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
298
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) UpperCAmelCase : Dict = { "configuration_speecht5": [ "SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP", "SpeechT5Config", "SpeechT5HifiGanConfig", ], "feature_extraction_speecht5": ["SpeechT5FeatureExtractor"], "processing_speecht5": ["SpeechT5Processor"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Tuple = ["SpeechT5Tokenizer"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Tuple = [ "SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST", "SpeechT5ForSpeechToText", "SpeechT5ForSpeechToSpeech", "SpeechT5ForTextToSpeech", "SpeechT5Model", "SpeechT5PreTrainedModel", "SpeechT5HifiGan", ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys UpperCAmelCase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
313
"""simple docstring""" import logging import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEncoder, BertModel, BertPreTrainedModel, ) UpperCAmelCase : Optional[int] = logging.getLogger(__name__) class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : int , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : Optional[Any]=None): """simple docstring""" lowercase_ = self.layer[current_layer](lowerCAmelCase_ , lowerCAmelCase_ , head_mask[current_layer]) lowercase_ = layer_outputs[0] return hidden_states @add_start_docstrings( "The bare Bert Model transformer with PABEE outputting raw hidden-states without any specific head on top." , __UpperCAmelCase , ) class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): def __init__( self : Any , lowerCAmelCase_ : Dict): """simple docstring""" super().__init__(lowerCAmelCase_) lowercase_ = BertEncoderWithPabee(lowerCAmelCase_) self.init_weights() lowercase_ = 0 lowercase_ = 0 lowercase_ = 0 lowercase_ = 0 def _UpperCAmelCase ( self : Dict , lowerCAmelCase_ : int): """simple docstring""" lowercase_ = threshold def _UpperCAmelCase ( self : str , lowerCAmelCase_ : Optional[Any]): """simple docstring""" lowercase_ = patience def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ = 0 lowercase_ = 0 def _UpperCAmelCase ( self : Optional[int]): """simple docstring""" lowercase_ = self.inference_layers_num / self.inference_instances_num lowercase_ = ( F'''*** Patience = {self.patience} Avg. Inference Layers = {avg_inf_layers:.2f} Speed Up =''' F''' {1 - avg_inf_layers / self.config.num_hidden_layers:.2f} ***''' ) print(lowerCAmelCase_) @add_start_docstrings_to_model_forward(lowerCAmelCase_) def _UpperCAmelCase ( self : Union[str, Any] , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : Optional[Any]=None , lowerCAmelCase_ : Dict=None , lowerCAmelCase_ : Dict=None , lowerCAmelCase_ : str=None , lowerCAmelCase_ : int=None , lowerCAmelCase_ : Optional[int]=None , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : int=None , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : List[Any]=False , ): """simple docstring""" if input_ids is not None and inputs_embeds is not None: raise ValueError("""You cannot specify both input_ids and inputs_embeds at the same time""") elif input_ids is not None: lowercase_ = input_ids.size() elif inputs_embeds is not None: lowercase_ = inputs_embeds.size()[:-1] else: raise ValueError("""You have to specify either input_ids or inputs_embeds""") lowercase_ = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: lowercase_ = torch.ones(lowerCAmelCase_ , device=lowerCAmelCase_) if token_type_ids is None: lowercase_ = torch.zeros(lowerCAmelCase_ , dtype=torch.long , device=lowerCAmelCase_) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. lowercase_ = self.get_extended_attention_mask(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if self.config.is_decoder and encoder_hidden_states is not None: lowercase_ , lowercase_ , lowercase_ = encoder_hidden_states.size() lowercase_ = (encoder_batch_size, encoder_sequence_length) if encoder_attention_mask is None: lowercase_ = torch.ones(lowerCAmelCase_ , device=lowerCAmelCase_) lowercase_ = self.invert_attention_mask(lowerCAmelCase_) else: lowercase_ = None # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] lowercase_ = self.get_head_mask(lowerCAmelCase_ , self.config.num_hidden_layers) lowercase_ = self.embeddings( input_ids=lowerCAmelCase_ , position_ids=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , inputs_embeds=lowerCAmelCase_) lowercase_ = embedding_output if self.training: lowercase_ = [] for i in range(self.config.num_hidden_layers): lowercase_ = self.encoder.adaptive_forward( lowerCAmelCase_ , current_layer=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , head_mask=lowerCAmelCase_) lowercase_ = self.pooler(lowerCAmelCase_) lowercase_ = output_layers[i](output_dropout(lowerCAmelCase_)) res.append(lowerCAmelCase_) elif self.patience == 0: # Use all layers for inference lowercase_ = self.encoder( lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , head_mask=lowerCAmelCase_ , encoder_hidden_states=lowerCAmelCase_ , encoder_attention_mask=lowerCAmelCase_ , ) lowercase_ = self.pooler(encoder_outputs[0]) lowercase_ = [output_layers[self.config.num_hidden_layers - 1](lowerCAmelCase_)] else: lowercase_ = 0 lowercase_ = None lowercase_ = 0 for i in range(self.config.num_hidden_layers): calculated_layer_num += 1 lowercase_ = self.encoder.adaptive_forward( lowerCAmelCase_ , current_layer=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , head_mask=lowerCAmelCase_) lowercase_ = self.pooler(lowerCAmelCase_) lowercase_ = output_layers[i](lowerCAmelCase_) if regression: lowercase_ = logits.detach() if patient_result is not None: lowercase_ = patient_result.detach() if (patient_result is not None) and torch.abs(patient_result - labels) < self.regression_threshold: patient_counter += 1 else: lowercase_ = 0 else: lowercase_ = logits.detach().argmax(dim=1) if patient_result is not None: lowercase_ = patient_result.detach().argmax(dim=1) if (patient_result is not None) and torch.all(labels.eq(lowerCAmelCase_)): patient_counter += 1 else: lowercase_ = 0 lowercase_ = logits if patient_counter == self.patience: break lowercase_ = [patient_result] self.inference_layers_num += calculated_layer_num self.inference_instances_num += 1 return res @add_start_docstrings( "Bert Model transformer with PABEE and a sequence classification/regression head on top (a linear layer on top of\n the pooled output) e.g. for GLUE tasks. " , __UpperCAmelCase , ) class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): def __init__( self : Optional[Any] , lowerCAmelCase_ : str): """simple docstring""" super().__init__(lowerCAmelCase_) lowercase_ = config.num_labels lowercase_ = BertModelWithPabee(lowerCAmelCase_) lowercase_ = nn.Dropout(config.hidden_dropout_prob) lowercase_ = nn.ModuleList( [nn.Linear(config.hidden_size , self.config.num_labels) for _ in range(config.num_hidden_layers)]) self.init_weights() @add_start_docstrings_to_model_forward(lowerCAmelCase_) def _UpperCAmelCase ( self : List[str] , lowerCAmelCase_ : Optional[Any]=None , lowerCAmelCase_ : Optional[int]=None , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : Union[str, Any]=None , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : List[Any]=None , lowerCAmelCase_ : List[str]=None , ): """simple docstring""" lowercase_ = self.bert( input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , position_ids=lowerCAmelCase_ , head_mask=lowerCAmelCase_ , inputs_embeds=lowerCAmelCase_ , output_dropout=self.dropout , output_layers=self.classifiers , regression=self.num_labels == 1 , ) lowercase_ = (logits[-1],) if labels is not None: lowercase_ = None lowercase_ = 0 for ix, logits_item in enumerate(lowerCAmelCase_): if self.num_labels == 1: # We are doing regression lowercase_ = MSELoss() lowercase_ = loss_fct(logits_item.view(-1) , labels.view(-1)) else: lowercase_ = CrossEntropyLoss() lowercase_ = loss_fct(logits_item.view(-1 , self.num_labels) , labels.view(-1)) if total_loss is None: lowercase_ = loss else: total_loss += loss * (ix + 1) total_weights += ix + 1 lowercase_ = (total_loss / total_weights,) + outputs return outputs
313
1
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _A = logging.get_logger(__name__) _A = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = "detr" UpperCAmelCase__ : Optional[int] = ["past_key_values"] UpperCAmelCase__ : Optional[int] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self , A_=True , A_=None , A_=3 , A_=100 , A_=6 , A_=2048 , A_=8 , A_=6 , A_=2048 , A_=8 , A_=0.0 , A_=0.0 , A_=True , A_="relu" , A_=256 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=1.0 , A_=False , A_="sine" , A_="resnet50" , A_=True , A_=False , A_=1 , A_=5 , A_=2 , A_=1 , A_=1 , A_=5 , A_=2 , A_=0.1 , **A_ , ) -> List[Any]: if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) __UpperCamelCase =CONFIG_MAPPING['resnet'](out_features=['stage4'] ) elif isinstance(A_ , A_ ): __UpperCamelCase =backbone_config.get('model_type' ) __UpperCamelCase =CONFIG_MAPPING[backbone_model_type] __UpperCamelCase =config_class.from_dict(A_ ) # set timm attributes to None __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =None, None, None __UpperCamelCase =use_timm_backbone __UpperCamelCase =backbone_config __UpperCamelCase =num_channels __UpperCamelCase =num_queries __UpperCamelCase =d_model __UpperCamelCase =encoder_ffn_dim __UpperCamelCase =encoder_layers __UpperCamelCase =encoder_attention_heads __UpperCamelCase =decoder_ffn_dim __UpperCamelCase =decoder_layers __UpperCamelCase =decoder_attention_heads __UpperCamelCase =dropout __UpperCamelCase =attention_dropout __UpperCamelCase =activation_dropout __UpperCamelCase =activation_function __UpperCamelCase =init_std __UpperCamelCase =init_xavier_std __UpperCamelCase =encoder_layerdrop __UpperCamelCase =decoder_layerdrop __UpperCamelCase =encoder_layers __UpperCamelCase =auxiliary_loss __UpperCamelCase =position_embedding_type __UpperCamelCase =backbone __UpperCamelCase =use_pretrained_backbone __UpperCamelCase =dilation # Hungarian matcher __UpperCamelCase =class_cost __UpperCamelCase =bbox_cost __UpperCamelCase =giou_cost # Loss coefficients __UpperCamelCase =mask_loss_coefficient __UpperCamelCase =dice_loss_coefficient __UpperCamelCase =bbox_loss_coefficient __UpperCamelCase =giou_loss_coefficient __UpperCamelCase =eos_coefficient super().__init__(is_encoder_decoder=A_ , **A_ ) @property def _a ( self ) -> int: return self.encoder_attention_heads @property def _a ( self ) -> int: return self.d_model @classmethod def _a ( cls , A_ , **A_ ) -> Tuple: return cls(backbone_config=A_ , **A_ ) def _a ( self ) -> Dict[str, any]: __UpperCamelCase =copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __UpperCamelCase =self.backbone_config.to_dict() __UpperCamelCase =self.__class__.model_type return output class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Dict = version.parse("1.11" ) @property def _a ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ] ) @property def _a ( self ) -> float: return 1E-5 @property def _a ( self ) -> int: return 12
62
"""simple docstring""" import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process snake_case__ : Dict = logging.getLogger(__name__) def _snake_case ( _snake_case : Any , _snake_case : Any ): return (preds == labels).mean() @dataclass class snake_case_: __UpperCamelCase = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) __UpperCamelCase = field( default=a__ , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''} ) __UpperCamelCase = field( default=a__ , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''} ) __UpperCamelCase = field( default=a__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) @dataclass class snake_case_: __UpperCamelCase = field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(processors.keys() )} ) __UpperCamelCase = field(metadata={'''help''': '''Should contain the data files for the task.'''} ) __UpperCamelCase = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) __UpperCamelCase = field( default=a__ , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def _snake_case ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowerCAmelCase, lowerCAmelCase, lowerCAmelCase : Optional[int] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f'''Output directory ({training_args.output_dir}) already exists and is not empty. Use''' ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , _snake_case ) # Set seed set_seed(training_args.seed ) try: lowerCAmelCase : Tuple = processors[data_args.task_name]() lowerCAmelCase : Any = processor.get_labels() lowerCAmelCase : Union[str, Any] = len(_snake_case ) except KeyError: raise ValueError('''Task not found: %s''' % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=_snake_case , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) lowerCAmelCase : List[str] = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=_snake_case , cache_dir=model_args.cache_dir , ) # Get datasets lowerCAmelCase : Dict = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_snake_case , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) lowerCAmelCase : Any = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_snake_case , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(_snake_case : EvalPrediction ) -> Dict: lowerCAmelCase : int = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(_snake_case , p.label_ids )} # Data collator lowerCAmelCase : List[Any] = DataCollatorWithPadding(_snake_case , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer lowerCAmelCase : Union[str, Any] = Trainer( model=_snake_case , args=_snake_case , train_dataset=_snake_case , eval_dataset=_snake_case , compute_metrics=_snake_case , data_collator=_snake_case , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowerCAmelCase : int = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowerCAmelCase : Any = trainer.evaluate() lowerCAmelCase : int = os.path.join(training_args.output_dir , '''eval_results.txt''' ) if trainer.is_world_master(): with open(_snake_case , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(''' %s = %s''' , _snake_case , _snake_case ) writer.write('''%s = %s\n''' % (key, value) ) results.update(_snake_case ) return results def _snake_case ( _snake_case : List[str] ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
60
0
'''simple docstring''' import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class SCREAMING_SNAKE_CASE (a__ ): def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[str] = tempfile.mkdtemp() __A : Tuple = 5 # Realm tok __A : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'test', 'question', 'this', 'is', 'the', 'first', 'second', 'third', 'fourth', 'fifth', 'record', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __A : Optional[Any] = os.path.join(self.tmpdirname , 'realm_tokenizer') os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase) __A : Dict = os.path.join(_UpperCAmelCase , VOCAB_FILES_NAMES['vocab_file']) with open(self.vocab_file , 'w' , encoding='utf-8') as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens])) __A : Dict = os.path.join(self.tmpdirname , 'realm_block_records') os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'realm_tokenizer')) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' shutil.rmtree(self.tmpdirname) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Tuple = RealmConfig(num_block_records=self.num_block_records) return config def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Any = Dataset.from_dict( { 'id': ['0', '1'], 'question': ['foo', 'bar'], 'answers': [['Foo', 'Bar'], ['Bar']], }) return dataset def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[int] = np.array( [ B'This is the first record', B'This is the second record', B'This is the third record', B'This is the fourth record', B'This is the fifth record', B'This is a longer longer longer record', ] , dtype=_UpperCAmelCase , ) return block_records def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[Any] = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[Any] = self.get_config() __A : Tuple = self.get_dummy_retriever() __A : Union[str, Any] = retriever.tokenizer __A : int = np.array([0, 3] , dtype='long') __A : Union[str, Any] = tokenizer(['Test question']).input_ids __A : List[Any] = tokenizer( ['the fourth'] , add_special_tokens=_UpperCAmelCase , return_token_type_ids=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , ).input_ids __A : int = config.reader_seq_len __A ,__A ,__A ,__A : Any = retriever( _UpperCAmelCase , _UpperCAmelCase , answer_ids=_UpperCAmelCase , max_length=_UpperCAmelCase , return_tensors='np') self.assertEqual(len(_UpperCAmelCase) , 2) self.assertEqual(len(_UpperCAmelCase) , 2) self.assertEqual(len(_UpperCAmelCase) , 2) self.assertEqual(concat_inputs.input_ids.shape , (2, 10)) self.assertEqual(concat_inputs.attention_mask.shape , (2, 10)) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 10)) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 10)) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0]) , ['[CLS]', 'test', 'question', '[SEP]', 'this', 'is', 'the', 'first', 'record', '[SEP]'] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1]) , ['[CLS]', 'test', 'question', '[SEP]', 'this', 'is', 'the', 'fourth', 'record', '[SEP]'] , ) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = self.get_config() __A : str = self.get_dummy_retriever() __A : int = retriever.tokenizer __A : Any = np.array([0, 3, 5] , dtype='long') __A : Tuple = tokenizer(['Test question']).input_ids __A : Dict = tokenizer( ['the fourth', 'longer longer'] , add_special_tokens=_UpperCAmelCase , return_token_type_ids=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , ).input_ids __A : int = config.reader_seq_len __A ,__A ,__A ,__A : Any = retriever( _UpperCAmelCase , _UpperCAmelCase , answer_ids=_UpperCAmelCase , max_length=_UpperCAmelCase , return_tensors='np') self.assertEqual([False, True, True] , _UpperCAmelCase) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , _UpperCAmelCase) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : str = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , 'realm_block_records')) # Test local path __A : Union[str, Any] = retriever.from_pretrained(os.path.join(self.tmpdirname , 'realm_block_records')) self.assertEqual(retriever.block_records[0] , B'This is the first record') # Test mocked remote path with patch('transformers.models.realm.retrieval_realm.hf_hub_download') as mock_hf_hub_download: __A : List[str] = os.path.join( os.path.join(self.tmpdirname , 'realm_block_records') , _REALM_BLOCK_RECORDS_FILENAME) __A : Dict = RealmRetriever.from_pretrained('google/realm-cc-news-pretrained-openqa') self.assertEqual(retriever.block_records[0] , B'This is the first record')
190
'''simple docstring''' from collections.abc import Iterable from typing import Any class SCREAMING_SNAKE_CASE : def __init__( self , _UpperCAmelCase = None): '''simple docstring''' __A : str = value __A : Node | None = None # Added in order to delete a node easier __A : Node | None = None __A : Node | None = None def __repr__( self): '''simple docstring''' from pprint import pformat if self.left is None and self.right is None: return str(self.value) return pformat({F'{self.value}': (self.left, self.right)} , indent=1) class SCREAMING_SNAKE_CASE : def __init__( self , _UpperCAmelCase = None): '''simple docstring''' __A : int = root def __str__( self): '''simple docstring''' return str(self.root) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' if new_children is not None: # reset its kids __A : Optional[int] = node.parent if node.parent is not None: # reset its parent if self.is_right(_UpperCAmelCase): # If it is the right children __A : int = new_children else: __A : Union[str, Any] = new_children else: __A : Tuple = new_children def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' if node.parent and node.parent.right: return node == node.parent.right return False def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return self.root is None def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' __A : Dict = Node(_UpperCAmelCase) # create a new Node if self.empty(): # if Tree is empty __A : Union[str, Any] = new_node # set its root else: # Tree is not empty __A : Dict = self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: __A : Tuple = new_node # We insert the new node in a leaf break else: __A : str = parent_node.left else: if parent_node.right is None: __A : List[str] = new_node break else: __A : int = parent_node.right __A : int = parent_node def SCREAMING_SNAKE_CASE ( self , *_UpperCAmelCase): '''simple docstring''' for value in values: self.__insert(_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' if self.empty(): raise IndexError('Warning: Tree is empty! please use another.') else: __A : Any = self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: __A : List[Any] = node.left if value < node.value else node.right return node def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase = None): '''simple docstring''' if node is None: if self.root is None: return None __A : str = self.root if not self.empty(): while node.right is not None: __A : Dict = node.right return node def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase = None): '''simple docstring''' if node is None: __A : Optional[Any] = self.root if self.root is None: return None if not self.empty(): __A : Optional[int] = self.root while node.left is not None: __A : Tuple = node.left return node def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' __A : Optional[Any] = self.search(_UpperCAmelCase) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(_UpperCAmelCase , _UpperCAmelCase) elif node.left is None: # Has only right children self.__reassign_nodes(_UpperCAmelCase , node.right) elif node.right is None: # Has only left children self.__reassign_nodes(_UpperCAmelCase , node.left) else: __A : str = self.get_max( node.left) # Gets the max value of the left branch self.remove(tmp_node.value) # type: ignore __A : List[Any] = ( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left) yield from self.preorder_traverse(node.right) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase=None): '''simple docstring''' if traversal_function is None: return self.preorder_traverse(self.root) else: return traversal_function(self.root) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' if node: self.inorder(_UpperCAmelCase , node.left) arr.append(node.value) self.inorder(_UpperCAmelCase , node.right) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : list[int] = [] self.inorder(_UpperCAmelCase , _UpperCAmelCase) # append all values to list using inorder traversal return arr[k - 1] def _lowerCAmelCase ( __snake_case : Node | None ) -> list[Node]: __A : Tuple = [] if curr_node is not None: __A : Optional[Any] = postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def _lowerCAmelCase ( ) -> None: __A : Optional[int] = (8, 3, 6, 1, 10, 14, 13, 4, 7) __A : str = BinarySearchTree() for i in testlist: t.insert(__snake_case ) # Prints all the elements of the list in order traversal print(__snake_case ) if t.search(6 ) is not None: print('The value 6 exists' ) else: print('The value 6 doesn\'t exist' ) if t.search(-1 ) is not None: print('The value -1 exists' ) else: print('The value -1 doesn\'t exist' ) if not t.empty(): print('Max Value: ' , t.get_max().value ) # type: ignore print('Min Value: ' , t.get_min().value ) # type: ignore for i in testlist: t.remove(__snake_case ) print(__snake_case ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
190
1
'''simple docstring''' from __future__ import annotations a_ : str = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] a_ : str = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def a_ ( __snake_case : list[float] ) -> list[float]: """simple docstring""" lowerCamelCase_ =[] lowerCamelCase_ =len(__snake_case ) for i in range(__snake_case ): lowerCamelCase_ =-1 for j in range(i + 1 , __snake_case ): if arr[i] < arr[j]: lowerCamelCase_ =arr[j] break result.append(__snake_case ) return result def a_ ( __snake_case : list[float] ) -> list[float]: """simple docstring""" lowerCamelCase_ =[] for i, outer in enumerate(__snake_case ): lowerCamelCase_ =-1 for inner in arr[i + 1 :]: if outer < inner: lowerCamelCase_ =inner break result.append(__snake_case ) return result def a_ ( __snake_case : list[float] ) -> list[float]: """simple docstring""" lowerCamelCase_ =len(__snake_case ) lowerCamelCase_ =[] lowerCamelCase_ =[-1] * arr_size for index in reversed(range(__snake_case ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: lowerCamelCase_ =stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) a_ : Dict = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
75
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer a__ : Optional[Any] = logging.get_logger(__name__) a__ : Union[str, Any] = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} a__ : Any = { 'vocab_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/vocab.txt' ), 'squeezebert/squeezebert-mnli': 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/vocab.txt', 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli': ( 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/tokenizer.json' ), }, } a__ : Optional[Any] = { 'squeezebert/squeezebert-uncased': 5_1_2, 'squeezebert/squeezebert-mnli': 5_1_2, 'squeezebert/squeezebert-mnli-headless': 5_1_2, } a__ : Optional[Any] = { 'squeezebert/squeezebert-uncased': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli-headless': {'do_lower_case': True}, } class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_INIT_CONFIGURATION __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = SqueezeBertTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=True , lowercase="[UNK]" , lowercase="[SEP]" , lowercase="[PAD]" , lowercase="[CLS]" , lowercase="[MASK]" , lowercase=True , lowercase=None , **lowercase , ) -> Tuple: super().__init__( lowercase , tokenizer_file=lowercase , do_lower_case=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , tokenize_chinese_chars=lowercase , strip_accents=lowercase , **lowercase , ) __UpperCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , lowercase ) != do_lower_case or normalizer_state.get("""strip_accents""" , lowercase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , lowercase ) != tokenize_chinese_chars ): __UpperCamelCase = getattr(lowercase , normalizer_state.pop("""type""" ) ) __UpperCamelCase = do_lower_case __UpperCamelCase = strip_accents __UpperCamelCase = tokenize_chinese_chars __UpperCamelCase = normalizer_class(**lowercase ) __UpperCamelCase = do_lower_case def __lowerCamelCase ( self , lowercase , lowercase=None ) -> Tuple: __UpperCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowerCamelCase ( self , lowercase , lowercase = None ) -> List[int]: __UpperCamelCase = [self.sep_token_id] __UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowerCamelCase ( self , lowercase , lowercase = None ) -> Tuple[str]: __UpperCamelCase = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase )
349
0
'''simple docstring''' import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model SCREAMING_SNAKE_CASE__ = '0.12' # assumed parallelism: 8 if is_torch_available(): import torch def lowercase__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None )-> List[Any]: if rng is None: UpperCamelCase = random.Random() UpperCamelCase = 1 for dim in shape: total_dims *= dim UpperCamelCase = [] for _ in range(_a ): values.append(rng.randint(0 , vocab_size - 1 ) ) UpperCamelCase = np.array(_a , dtype=jnp.intaa ).reshape(_a ) return output def lowercase__ ( __UpperCamelCase , __UpperCamelCase=None )-> Any: UpperCamelCase = ids_tensor(_a , vocab_size=2 , rng=_a ) # make sure that at least one token is attended to for each batch UpperCamelCase = 1 return attn_mask @require_flax class a_ : lowercase = None lowercase = () def A__ ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 UpperCamelCase = 2 UpperCamelCase = inputs["""input_ids"""].shape[-1] // 2 UpperCamelCase = inputs["""input_ids"""][:max_batch_size, :sequence_length] UpperCamelCase = jnp.ones_like(lowerCamelCase_ ) UpperCamelCase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens UpperCamelCase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` UpperCamelCase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def A__ ( self ) -> Optional[int]: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length UpperCamelCase = 0 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning UpperCamelCase = getattr(lowerCamelCase_ , lowerCamelCase_ ) UpperCamelCase = pt_model_class(lowerCamelCase_ ).eval() UpperCamelCase = load_flax_weights_in_pytorch_model(lowerCamelCase_ , flax_model.params ) UpperCamelCase = flax_model.generate(lowerCamelCase_ ).sequences UpperCamelCase = pt_model.generate(torch.tensor(lowerCamelCase_ , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: UpperCamelCase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def A__ ( self ) -> int: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> str: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = True UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> Optional[int]: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length UpperCamelCase = 2 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = False UpperCamelCase = max_length UpperCamelCase = 2 UpperCamelCase = 2 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def A__ ( self ) -> int: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = True UpperCamelCase = max_length UpperCamelCase = 0.8 UpperCamelCase = 10 UpperCamelCase = 0.3 UpperCamelCase = 1 UpperCamelCase = 8 UpperCamelCase = 9 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> str: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = max_length UpperCamelCase = 1 UpperCamelCase = 8 UpperCamelCase = 9 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() UpperCamelCase = max_length UpperCamelCase = 2 UpperCamelCase = 1 UpperCamelCase = 8 UpperCamelCase = 9 for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> Any: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() # pad attention mask on the left UpperCamelCase = attention_mask.at[(0, 0)].set(0 ) UpperCamelCase = False UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ , attention_mask=lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ , attention_mask=lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() # pad attention mask on the left UpperCamelCase = attention_mask.at[(0, 0)].set(0 ) UpperCamelCase = True UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ , attention_mask=lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ , attention_mask=lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def A__ ( self ) -> List[Any]: """simple docstring""" UpperCamelCase = self._get_input_ids_and_config() # pad attention mask on the left UpperCamelCase = attention_mask.at[(0, 0)].set(0 ) UpperCamelCase = 2 UpperCamelCase = max_length for model_class in self.all_generative_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model.generate(lowerCamelCase_ , attention_mask=lowerCamelCase_ ).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCamelCase_ ) UpperCamelCase = jit(model.generate ) UpperCamelCase = jit_generate(lowerCamelCase_ , attention_mask=lowerCamelCase_ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class a_ ( unittest.TestCase ): def A__ ( self ) -> Optional[Any]: """simple docstring""" UpperCamelCase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-bert""" ) UpperCamelCase = FlaxAutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) UpperCamelCase = """Hello world""" UpperCamelCase = tokenizer(lowerCamelCase_ , return_tensors="""np""" ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(lowerCamelCase_ , """do_samples""" ): model.generate(lowerCamelCase_ , do_samples=lowerCamelCase_ ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(lowerCamelCase_ , """foo""" ): UpperCamelCase = {"""foo""": """bar"""} model.generate(lowerCamelCase_ , **lowerCamelCase_ )
365
'''simple docstring''' import functools import logging import os import sys import threading from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional import huggingface_hub.utils as hf_hub_utils from tqdm import auto as tqdm_lib SCREAMING_SNAKE_CASE__ = threading.Lock() SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = { 'debug': logging.DEBUG, 'info': logging.INFO, 'warning': logging.WARNING, 'error': logging.ERROR, 'critical': logging.CRITICAL, } SCREAMING_SNAKE_CASE__ = logging.WARNING SCREAMING_SNAKE_CASE__ = True def lowercase__ ( )-> Optional[int]: UpperCamelCase = os.getenv("""TRANSFORMERS_VERBOSITY""" , __UpperCamelCase ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F"Unknown option TRANSFORMERS_VERBOSITY={env_level_str}, " F"has to be one of: { ', '.join(log_levels.keys() ) }" ) return _default_log_level def lowercase__ ( )-> str: return __name__.split(""".""" )[0] def lowercase__ ( )-> logging.Logger: return logging.getLogger(_get_library_name() ) def lowercase__ ( )-> None: global _default_handler with _lock: if _default_handler: # This library has already configured the library root logger. return UpperCamelCase = logging.StreamHandler() # Set sys.stderr as stream. UpperCamelCase = sys.stderr.flush # Apply our default configuration to the library root logger. UpperCamelCase = _get_library_root_logger() library_root_logger.addHandler(_default_handler ) library_root_logger.setLevel(_get_default_logging_level() ) UpperCamelCase = False def lowercase__ ( )-> None: global _default_handler with _lock: if not _default_handler: return UpperCamelCase = _get_library_root_logger() library_root_logger.removeHandler(_default_handler ) library_root_logger.setLevel(logging.NOTSET ) UpperCamelCase = None def lowercase__ ( )-> Tuple: return log_levels def lowercase__ ( __UpperCamelCase = None )-> logging.Logger: if name is None: UpperCamelCase = _get_library_name() _configure_library_root_logger() return logging.getLogger(__UpperCamelCase ) def lowercase__ ( )-> int: _configure_library_root_logger() return _get_library_root_logger().getEffectiveLevel() def lowercase__ ( __UpperCamelCase )-> None: _configure_library_root_logger() _get_library_root_logger().setLevel(__UpperCamelCase ) def lowercase__ ( )-> Tuple: return set_verbosity(__UpperCamelCase ) def lowercase__ ( )-> Union[str, Any]: return set_verbosity(__UpperCamelCase ) def lowercase__ ( )-> Optional[int]: return set_verbosity(__UpperCamelCase ) def lowercase__ ( )-> Tuple: return set_verbosity(__UpperCamelCase ) def lowercase__ ( )-> None: _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().removeHandler(_default_handler ) def lowercase__ ( )-> None: _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().addHandler(_default_handler ) def lowercase__ ( __UpperCamelCase )-> None: _configure_library_root_logger() assert handler is not None _get_library_root_logger().addHandler(__UpperCamelCase ) def lowercase__ ( __UpperCamelCase )-> None: _configure_library_root_logger() assert handler is not None and handler not in _get_library_root_logger().handlers _get_library_root_logger().removeHandler(__UpperCamelCase ) def lowercase__ ( )-> None: _configure_library_root_logger() UpperCamelCase = False def lowercase__ ( )-> None: _configure_library_root_logger() UpperCamelCase = True def lowercase__ ( )-> None: UpperCamelCase = _get_library_root_logger().handlers for handler in handlers: UpperCamelCase = logging.Formatter("""[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s""" ) handler.setFormatter(__UpperCamelCase ) def lowercase__ ( )-> None: UpperCamelCase = _get_library_root_logger().handlers for handler in handlers: handler.setFormatter(__UpperCamelCase ) def lowercase__ ( self , *__UpperCamelCase , **__UpperCamelCase )-> Tuple: UpperCamelCase = os.getenv("""TRANSFORMERS_NO_ADVISORY_WARNINGS""" , __UpperCamelCase ) if no_advisory_warnings: return self.warning(*__UpperCamelCase , **__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = warning_advice @functools.lru_cache(__UpperCamelCase ) def lowercase__ ( self , *__UpperCamelCase , **__UpperCamelCase )-> Optional[Any]: self.warning(*__UpperCamelCase , **__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = warning_once class a_ : def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Optional[int]: # pylint: disable=unused-argument """simple docstring""" UpperCamelCase = args[0] if args else None def __iter__( self ) -> List[Any]: """simple docstring""" return iter(self._iterator ) def __getattr__( self , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" def empty_fn(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): # pylint: disable=unused-argument return return empty_fn def __enter__( self ) -> Dict: """simple docstring""" return self def __exit__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" return class a_ : def __call__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" if _tqdm_active: return tqdm_lib.tqdm(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) else: return EmptyTqdm(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def A__ ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" UpperCamelCase = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def A__ ( self ) -> Union[str, Any]: """simple docstring""" if _tqdm_active: return tqdm_lib.tqdm.get_lock() SCREAMING_SNAKE_CASE__ = _tqdm_cls() def lowercase__ ( )-> bool: global _tqdm_active return bool(_tqdm_active ) def lowercase__ ( )-> Optional[Any]: global _tqdm_active UpperCamelCase = True hf_hub_utils.enable_progress_bars() def lowercase__ ( )-> str: global _tqdm_active UpperCamelCase = False hf_hub_utils.disable_progress_bars()
183
0
from math import factorial class __snake_case : def __init__( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : List[str] = real if isinstance(snake_case ,snake_case ): lowercase : Any = [1] * rank else: lowercase : Optional[Any] = rank def __repr__( self ): '''simple docstring''' return ( f"{self.real}+" f"{'+'.join(str(snake_case )+'E'+str(n+1 )for n,dual in enumerate(self.duals ) )}" ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Dict = self.duals.copy() while cur[-1] == 0: cur.pop(-1 ) return Dual(self.real ,snake_case ) def __add__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): return Dual(self.real + other ,self.duals ) lowercase : Optional[int] = self.duals.copy() lowercase : Optional[Any] = other.duals.copy() if len(snake_case ) > len(snake_case ): o_dual.extend([1] * (len(snake_case ) - len(snake_case )) ) elif len(snake_case ) < len(snake_case ): s_dual.extend([1] * (len(snake_case ) - len(snake_case )) ) lowercase : Dict = [] for i in range(len(snake_case ) ): new_duals.append(s_dual[i] + o_dual[i] ) return Dual(self.real + other.real ,snake_case ) _a : Union[str, Any]= __add__ def __sub__( self ,snake_case ): '''simple docstring''' return self + other * -1 def __mul__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): lowercase : str = [] for i in self.duals: new_duals.append(i * other ) return Dual(self.real * other ,snake_case ) lowercase : Union[str, Any] = [0] * (len(self.duals ) + len(other.duals ) + 1) for i, item in enumerate(self.duals ): for j, jtem in enumerate(other.duals ): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals ) ): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals ) ): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real ,snake_case ) _a : int= __mul__ def __truediv__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): lowercase : Union[str, Any] = [] for i in self.duals: new_duals.append(i / other ) return Dual(self.real / other ,snake_case ) raise ValueError def __floordiv__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): lowercase : List[Any] = [] for i in self.duals: new_duals.append(i // other ) return Dual(self.real // other ,snake_case ) raise ValueError def __pow__( self ,snake_case ): '''simple docstring''' if n < 0 or isinstance(snake_case ,snake_case ): raise ValueError("""power must be a positive integer""" ) if n == 0: return 1 if n == 1: return self lowercase : str = self for _ in range(n - 1 ): x *= self return x def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Dict: if not callable(SCREAMING_SNAKE_CASE__ ): raise ValueError("""differentiate() requires a function as input for func""" ) if not isinstance(SCREAMING_SNAKE_CASE__ , (float, int) ): raise ValueError("""differentiate() requires a float as input for position""" ) if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise ValueError("""differentiate() requires an int as input for order""" ) lowercase : List[Any] = Dual(SCREAMING_SNAKE_CASE__ , 1 ) lowercase : List[Any] = func(SCREAMING_SNAKE_CASE__ ) if order == 0: return result.real return result.duals[order - 1] * factorial(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": import doctest doctest.testmod() def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: return y**2 * y**4 print(differentiate(f, 9, 2))
20
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers UpperCamelCase__: Optional[int] = "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
23
0
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCamelCase : '''simple docstring''' def __init__( self : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any]=2 , _lowerCAmelCase : Dict=8 , _lowerCAmelCase : List[Any]=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Tuple=True , _lowerCAmelCase : Any=True , _lowerCAmelCase : Dict=9_9 , _lowerCAmelCase : List[Any]=1_6 , _lowerCAmelCase : Union[str, Any]=5 , _lowerCAmelCase : List[Any]=2 , _lowerCAmelCase : Union[str, Any]=3_6 , _lowerCAmelCase : Union[str, Any]="gelu" , _lowerCAmelCase : Any=0.0 , _lowerCAmelCase : Union[str, Any]=0.0 , _lowerCAmelCase : Dict=5_1_2 , _lowerCAmelCase : Union[str, Any]=1_6 , _lowerCAmelCase : Optional[int]=2 , _lowerCAmelCase : Tuple=0.02 , _lowerCAmelCase : Optional[int]=3 , _lowerCAmelCase : str=4 , _lowerCAmelCase : List[Any]=None , ): '''simple docstring''' __lowercase =parent __lowercase =batch_size __lowercase =seq_length __lowercase =is_training __lowercase =use_input_mask __lowercase =use_token_type_ids __lowercase =use_labels __lowercase =vocab_size __lowercase =hidden_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =intermediate_size __lowercase =hidden_act __lowercase =hidden_dropout_prob __lowercase =attention_probs_dropout_prob __lowercase =max_position_embeddings __lowercase =type_vocab_size __lowercase =type_sequence_label_size __lowercase =initializer_range __lowercase =num_labels __lowercase =num_choices __lowercase =scope def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __lowercase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __lowercase =None if self.use_input_mask: __lowercase =random_attention_mask([self.batch_size, self.seq_length]) __lowercase =None if self.use_token_type_ids: __lowercase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __lowercase =None __lowercase =None __lowercase =None if self.use_labels: __lowercase =ids_tensor([self.batch_size] , self.type_sequence_label_size) __lowercase =ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __lowercase =ids_tensor([self.batch_size] , self.num_choices) __lowercase =self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCamelCase ( self : Any): '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def __lowerCamelCase ( self : Tuple): '''simple docstring''' __lowercase =self.get_config() __lowercase =3_0_0 return config def __lowerCamelCase ( self : List[str]): '''simple docstring''' ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) =self.prepare_config_and_inputs() __lowercase =True __lowercase =floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) __lowercase =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowerCamelCase ( self : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any , _lowerCAmelCase : Any , _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple): '''simple docstring''' __lowercase =MraModel(config=_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase) __lowercase =model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase) __lowercase =model(_lowerCAmelCase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def __lowerCamelCase ( self : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , ): '''simple docstring''' __lowercase =True __lowercase =MraModel(_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , ) __lowercase =model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , ) __lowercase =model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def __lowerCamelCase ( self : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any]): '''simple docstring''' __lowercase =MraForMaskedLM(config=_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def __lowerCamelCase ( self : List[str] , _lowerCAmelCase : str , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any]): '''simple docstring''' __lowercase =MraForQuestionAnswering(config=_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def __lowerCamelCase ( self : List[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]): '''simple docstring''' __lowercase =self.num_labels __lowercase =MraForSequenceClassification(_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def __lowerCamelCase ( self : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any]): '''simple docstring''' __lowercase =self.num_labels __lowercase =MraForTokenClassification(config=_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def __lowerCamelCase ( self : int , _lowerCAmelCase : str , _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int]): '''simple docstring''' __lowercase =self.num_choices __lowercase =MraForMultipleChoice(config=_lowerCAmelCase) model.to(_lowerCAmelCase) model.eval() __lowercase =input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowercase =token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowercase =input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __lowercase =model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def __lowerCamelCase ( self : Any): '''simple docstring''' __lowercase =self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) =config_and_inputs __lowercase ={'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class _UpperCamelCase ( A , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = () def __lowerCamelCase ( self : Tuple): '''simple docstring''' __lowercase =MraModelTester(self) __lowercase =ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=3_7) def __lowerCamelCase ( self : Tuple): '''simple docstring''' self.config_tester.run_common_tests() def __lowerCamelCase ( self : List[Any]): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase) def __lowerCamelCase ( self : Any): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowercase =type self.model_tester.create_and_check_model(*_lowerCAmelCase) def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCAmelCase) def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_lowerCAmelCase) def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCAmelCase) def __lowerCamelCase ( self : List[str]): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCAmelCase) def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCAmelCase) @slow def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase =MraModel.from_pretrained(_lowerCAmelCase) self.assertIsNotNone(_lowerCAmelCase) @unittest.skip(reason='MRA does not output attentions') def __lowerCamelCase ( self : Dict): '''simple docstring''' return @require_torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def __lowerCamelCase ( self : int): '''simple docstring''' __lowercase =MraModel.from_pretrained('uw-madison/mra-base-512-4') __lowercase =torch.arange(2_5_6).unsqueeze(0) with torch.no_grad(): __lowercase =model(_lowerCAmelCase)[0] __lowercase =torch.Size((1, 2_5_6, 7_6_8)) self.assertEqual(output.shape , _lowerCAmelCase) __lowercase =torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=1e-4)) @slow def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =MraForMaskedLM.from_pretrained('uw-madison/mra-base-512-4') __lowercase =torch.arange(2_5_6).unsqueeze(0) with torch.no_grad(): __lowercase =model(_lowerCAmelCase)[0] __lowercase =5_0_2_6_5 __lowercase =torch.Size((1, 2_5_6, vocab_size)) self.assertEqual(output.shape , _lowerCAmelCase) __lowercase =torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=1e-4)) @slow def __lowerCamelCase ( self : int): '''simple docstring''' __lowercase =MraForMaskedLM.from_pretrained('uw-madison/mra-base-4096-8-d3') __lowercase =torch.arange(4_0_9_6).unsqueeze(0) with torch.no_grad(): __lowercase =model(_lowerCAmelCase)[0] __lowercase =5_0_2_6_5 __lowercase =torch.Size((1, 4_0_9_6, vocab_size)) self.assertEqual(output.shape , _lowerCAmelCase) __lowercase =torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=1e-4))
370
'''simple docstring''' import functools def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =len(_lowerCAmelCase ) __lowercase =len(_lowerCAmelCase ) @functools.cache def min_distance(_lowerCAmelCase , _lowerCAmelCase ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa __lowercase =int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 , _lowerCAmelCase ) , 1 + min_distance(_lowerCAmelCase , indexa + 1 ) , diff + min_distance(indexa + 1 , indexa + 1 ) , ) return min_distance(0 , 0 ) if __name__ == "__main__": import doctest doctest.testmod()
48
0
"""simple docstring""" import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel __lowerCAmelCase : List[Any] ={ """text_branch""": """text_model""", """audio_branch""": """audio_model.audio_encoder""", """attn""": """attention.self""", """self.proj""": """output.dense""", """attention.self_mask""": """attn_mask""", """mlp.fc1""": """intermediate.dense""", """mlp.fc2""": """output.dense""", """norm1""": """layernorm_before""", """norm2""": """layernorm_after""", """bn0""": """batch_norm""", } __lowerCAmelCase : List[str] =AutoFeatureExtractor.from_pretrained("""laion/clap-htsat-unfused""", truncation="""rand_trunc""") def UpperCAmelCase__ ( lowerCAmelCase__ :List[str] , lowerCAmelCase__ :List[str]=False ) -> Optional[Any]: '''simple docstring''' lowercase , lowercase = create_model( """HTSAT-tiny""" , """roberta""" , lowerCAmelCase__ , precision="""fp32""" , device="""cuda:0""" if torch.cuda.is_available() else """cpu""" , enable_fusion=lowerCAmelCase__ , fusion_type="""aff_2d""" if enable_fusion else None , ) return model, model_cfg def UpperCAmelCase__ ( lowerCAmelCase__ :Optional[Any] ) -> Dict: '''simple docstring''' lowercase = {} lowercase = R""".*sequential.(\d+).*""" lowercase = R""".*_projection.(\d+).*""" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowercase = key.replace(lowerCAmelCase__ , lowerCAmelCase__ ) if re.match(lowerCAmelCase__ , lowerCAmelCase__ ): # replace sequential layers with list lowercase = re.match(lowerCAmelCase__ , lowerCAmelCase__ ).group(1 ) lowercase = key.replace(f'sequential.{sequential_layer}.' , f'layers.{int(lowerCAmelCase__ )//3}.linear.' ) elif re.match(lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = int(re.match(lowerCAmelCase__ , lowerCAmelCase__ ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... lowercase = 1 if projecton_layer == 0 else 2 lowercase = key.replace(f'_projection.{projecton_layer}.' , f'_projection.linear{transformers_projection_layer}.' ) if "audio" and "qkv" in key: # split qkv into query key and value lowercase = value lowercase = mixed_qkv.size(0 ) // 3 lowercase = mixed_qkv[:qkv_dim] lowercase = mixed_qkv[qkv_dim : qkv_dim * 2] lowercase = mixed_qkv[qkv_dim * 2 :] lowercase = query_layer lowercase = key_layer lowercase = value_layer else: lowercase = value return model_state_dict def UpperCAmelCase__ ( lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :int , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[int]=False ) -> int: '''simple docstring''' lowercase , lowercase = init_clap(lowerCAmelCase__ , enable_fusion=lowerCAmelCase__ ) clap_model.eval() lowercase = clap_model.state_dict() lowercase = rename_state_dict(lowerCAmelCase__ ) lowercase = ClapConfig() lowercase = enable_fusion lowercase = ClapModel(lowerCAmelCase__ ) # ignore the spectrogram embedding layer model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) model.save_pretrained(lowerCAmelCase__ ) transformers_config.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": __lowerCAmelCase : Optional[int] =argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument("""--enable_fusion""", action="""store_true""", help="""Whether to enable fusion or not""") __lowerCAmelCase : str =parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
197
"""simple docstring""" import argparse from pathlib import Path import torch from packaging import version from torch.onnx import export from diffusers import AutoencoderKL __lowerCAmelCase : Dict =version.parse(version.parse(torch.__version__).base_version) < version.parse("""1.11""") def UpperCAmelCase__ ( lowerCAmelCase__ :int , lowerCAmelCase__ :tuple , lowerCAmelCase__ :Path , lowerCAmelCase__ :str , lowerCAmelCase__ :int , lowerCAmelCase__ :Union[str, Any] , lowerCAmelCase__ :Any , lowerCAmelCase__ :List[str]=False , ) -> Union[str, Any]: '''simple docstring''' output_path.parent.mkdir(parents=lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , use_external_data_format=lowerCAmelCase__ , enable_onnx_checker=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) else: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) @torch.no_grad() def UpperCAmelCase__ ( lowerCAmelCase__ :str , lowerCAmelCase__ :str , lowerCAmelCase__ :int , lowerCAmelCase__ :bool = False ) -> str: '''simple docstring''' lowercase = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): lowercase = """cuda""" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: lowercase = """cpu""" lowercase = Path(lowerCAmelCase__ ) # VAE DECODER lowercase = AutoencoderKL.from_pretrained(model_path + """/vae""" ) lowercase = vae_decoder.config.latent_channels # forward only through the decoder part lowercase = vae_decoder.decode onnx_export( lowerCAmelCase__ , model_args=( torch.randn(1 , lowerCAmelCase__ , 2_5 , 2_5 ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase__ , ) del vae_decoder if __name__ == "__main__": __lowerCAmelCase : Tuple =argparse.ArgumentParser() parser.add_argument( """--model_path""", type=str, required=True, help="""Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).""", ) parser.add_argument("""--output_path""", type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--opset""", default=1_4, type=int, help="""The version of the ONNX operator set to use.""", ) parser.add_argument("""--fp16""", action="""store_true""", default=False, help="""Export the models in `float16` mode""") __lowerCAmelCase : Dict =parser.parse_args() print(args.output_path) convert_models(args.model_path, args.output_path, args.opset, args.fpaa) print("""SD: Done: ONNX""")
197
1
"""simple docstring""" import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): lowerCamelCase_ : Optional[Any] = """pt""" elif is_tf_available(): lowerCamelCase_ : Optional[int] = """tf""" else: lowerCamelCase_ : Any = """jax""" class __A ( _SCREAMING_SNAKE_CASE, unittest.TestCase ): """simple docstring""" __lowerCAmelCase = PerceiverTokenizer __lowerCAmelCase = False def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: super().setUp() a =PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def SCREAMING_SNAKE_CASE ( self , **__A ) -> PerceiverTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A=False , __A=20 , __A=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for Perceiver because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. a =[] for i in range(len(__A ) ): try: a =tokenizer.decode([i] , clean_up_tokenization_spaces=__A ) except UnicodeDecodeError: pass toks.append((i, tok) ) a =list(filter(lambda __A : re.match(r'''^[ a-zA-Z]+$''' , t[1] ) , __A ) ) a =list(filter(lambda __A : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=__A ) , __A ) ) if max_length is not None and len(__A ) > max_length: a =toks[:max_length] if min_length is not None and len(__A ) < min_length and len(__A ) > 0: while len(__A ) < min_length: a =toks + toks # toks_str = [t[1] for t in toks] a =[t[0] for t in toks] # Ensure consistency a =tokenizer.decode(__A , clean_up_tokenization_spaces=__A ) if " " not in output_txt and len(__A ) > 1: a =( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=__A ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=__A ) ) if with_prefix_space: a =''' ''' + output_txt a =tokenizer.encode(__A , add_special_tokens=__A ) return output_txt, output_ids def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: a =self.perceiver_tokenizer a ='''Unicode €.''' a =tokenizer(__A ) a =[4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded['''input_ids'''] , __A ) # decoding a =tokenizer.decode(__A ) self.assertEqual(__A , '''[CLS]Unicode €.[SEP]''' ) a =tokenizer('''e è é ê ë''' ) a =[4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded['''input_ids'''] , __A ) # decoding a =tokenizer.decode(__A ) self.assertEqual(__A , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def SCREAMING_SNAKE_CASE ( self ) -> Any: a =self.perceiver_tokenizer a =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off a =[4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on a =tokenizer(__A , padding=__A , return_tensors=__A ) self.assertIsInstance(__A , __A ) if FRAMEWORK != "jax": a =list(batch.input_ids.numpy()[0] ) else: a =list(batch.input_ids.tolist()[0] ) self.assertListEqual(__A , __A ) self.assertEqual((2, 38) , batch.input_ids.shape ) self.assertEqual((2, 38) , batch.attention_mask.shape ) def SCREAMING_SNAKE_CASE ( self ) -> List[str]: a =self.perceiver_tokenizer a =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] a =tokenizer(__A , padding=__A , return_tensors=__A ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , __A ) self.assertIn('''attention_mask''' , __A ) self.assertNotIn('''decoder_input_ids''' , __A ) self.assertNotIn('''decoder_attention_mask''' , __A ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: a =self.perceiver_tokenizer a =[ '''Summary of the text.''', '''Another summary.''', ] a =tokenizer( text_target=__A , max_length=32 , padding='''max_length''' , truncation=__A , return_tensors=__A ) self.assertEqual(32 , targets['''input_ids'''].shape[1] ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: # safety check on max_len default value so we are sure the test works a =self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test a =self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc a =tempfile.mkdtemp() a =''' He is very happy, UNwant\u00E9d,running''' a =tokenizer.encode(__A , add_special_tokens=__A ) tokenizer.save_pretrained(__A ) a =tokenizer.__class__.from_pretrained(__A ) a =after_tokenizer.encode(__A , add_special_tokens=__A ) self.assertListEqual(__A , __A ) shutil.rmtree(__A ) a =self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc a =tempfile.mkdtemp() a =''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) a =tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) a =tokenizer.encode(__A , add_special_tokens=__A ) tokenizer.save_pretrained(__A ) a =tokenizer.__class__.from_pretrained(__A ) a =after_tokenizer.encode(__A , add_special_tokens=__A ) self.assertListEqual(__A , __A ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) a =tokenizer.__class__.from_pretrained(__A , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(__A ) def SCREAMING_SNAKE_CASE ( self ) -> str: a =[] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(__A ) with open(os.path.join(__A , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: a =json.load(__A ) with open(os.path.join(__A , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: a =json.load(__A ) a =[f'''<extra_id_{i}>''' for i in range(125 )] a =added_tokens_extra_ids + [ '''an_additional_special_token''' ] a =added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(__A , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__A , __A ) with open(os.path.join(__A , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(__A , __A ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files a =tokenizer_class.from_pretrained( __A , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained a =added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=__A )] a =tokenizer_class.from_pretrained( __A , additional_special_tokens=__A , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ) , '''�''' ) def SCREAMING_SNAKE_CASE ( self ) -> Tuple: pass def SCREAMING_SNAKE_CASE ( self ) -> str: pass def SCREAMING_SNAKE_CASE ( self ) -> Tuple: pass def SCREAMING_SNAKE_CASE ( self ) -> List[str]: pass def SCREAMING_SNAKE_CASE ( self ) -> int: # The default common tokenizer tests uses invalid tokens for Perceiver that can only accept one-character # strings and special added tokens as tokens a =self.get_tokenizers(fast=__A , do_lower_case=__A ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): a =['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] a =tokenizer.convert_tokens_to_string(__A ) self.assertIsInstance(__A , __A )
362
"""simple docstring""" def _A ( lowercase = 2_00_00_00 ): """simple docstring""" a =[0 for i in range(n + 1 )] a =1 a =1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowercase ): a =1 a =0 for i in range(lowercase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'{solution() = }')
215
0
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex SCREAMING_SNAKE_CASE :int = logging.getLogger(__name__) class UpperCAmelCase : '''simple docstring''' def __init__( self : int ): __A = False def UpperCamelCase_ ( self : Optional[Any] ,A : Optional[Any] ,A : str ,A : Optional[Any] ,A : Union[str, Any] ): if not self.initialized: __A = RagRetriever( A ,question_encoder_tokenizer=A ,generator_tokenizer=A ,index=A ,init_retrieval=A ,) __A = True def UpperCamelCase_ ( self : List[Any] ): self.retriever.index.init_index() def UpperCamelCase_ ( self : List[Any] ,A : Optional[int] ,A : Dict ): __A , __A = self.retriever._main_retrieve(A ,A ) return doc_ids, retrieved_doc_embeds class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Tuple ,A : int ,A : Union[str, Any] ,A : List[Any] ,A : Tuple ,A : List[str]=None ): if index is not None and index.is_initialized() and len(A ) > 0: raise ValueError( "When using Ray for distributed fine-tuning, " "you'll need to provide the paths instead, " "as the dataset and the index are loaded " "separately. More info in examples/rag/use_own_knowledge_dataset.py " ) super().__init__( A ,question_encoder_tokenizer=A ,generator_tokenizer=A ,index=A ,init_retrieval=A ,) __A = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(A ,A ,A ,A ) for worker in self.retrieval_workers ] ) def UpperCamelCase_ ( self : str ): logger.info("initializing retrieval" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def UpperCamelCase_ ( self : Optional[Any] ,A : List[Any] ,A : List[Any] ): if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. __A = self.retrieval_workers[random.randint(0 ,len(self.retrieval_workers ) - 1 )] __A , __A = ray.get(random_worker.retrieve.remote(A ,A ) ) else: __A , __A = self._main_retrieve(A ,A ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(A ) @classmethod def UpperCamelCase_ ( cls : str ,A : List[Any] ,A : int=None ,**A : Optional[Any] ): return super(A ,cls ).get_tokenizers(A ,A ,**A ) @classmethod def UpperCamelCase_ ( cls : Any ,A : List[str] ,A : Optional[Any] ,A : List[str]=None ,**A : List[str] ): __A = kwargs.pop("config" ,A ) or RagConfig.from_pretrained(A ,**A ) __A = RagTokenizer.from_pretrained(A ,config=A ) __A = rag_tokenizer.question_encoder __A = rag_tokenizer.generator if indexed_dataset is not None: __A = "custom" __A = CustomHFIndex(config.retrieval_vector_size ,A ) else: __A = cls._build_index(A ) return cls( A ,question_encoder_tokenizer=A ,generator_tokenizer=A ,retrieval_workers=A ,index=A ,)
15
def UpperCAmelCase ( a_ ) -> list: """simple docstring""" if len(a_ ) <= 1: return [tuple(a_ )] __A = [] def generate(a_ , a_ ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , a_ ) for i in range(k - 1 ): if k % 2 == 0: # k is even __A , __A = arr[k - 1], arr[i] else: # k is odd __A , __A = arr[k - 1], arr[0] generate(k - 1 , a_ ) generate(len(a_ ) , a_ ) return res if __name__ == "__main__": SCREAMING_SNAKE_CASE :int = input('Enter numbers separated by a comma:\n').strip() SCREAMING_SNAKE_CASE :Dict = [int(item) for item in user_input.split(',')] print(heaps(arr))
15
1
import os import pytest from attr import dataclass _A = "us-east-1" # defaults region @dataclass class lowerCamelCase : UpperCAmelCase__ : str UpperCAmelCase__ : Optional[int] = "arn:aws:iam::558105141721:role/sagemaker_execution_role" UpperCAmelCase__ : List[Any] = { "task_name": "mnli", "per_device_train_batch_size": 16, "per_device_eval_batch_size": 16, "do_train": True, "do_eval": True, "do_predict": True, "output_dir": "/opt/ml/model", "overwrite_output_dir": True, "max_steps": 5_00, "save_steps": 55_00, } UpperCAmelCase__ : str = {**hyperparameters, "max_steps": 10_00} @property def UpperCAmelCase(self : Union[str, Any] ) -> str: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def UpperCAmelCase(self : Optional[int] ) -> str: return f'{self.framework}-transfromers-test' @property def UpperCAmelCase(self : List[str] ) -> str: return f'./tests/sagemaker/scripts/{self.framework}' @property def UpperCAmelCase(self : Dict ) -> str: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="class" ) def lowercase_ ( A__ ) -> int: """simple docstring""" snake_case = SageMakerTestEnvironment(framework=request.cls.framework )
137
from __future__ import annotations def lowercase_ ( A__ , A__ , A__ ) -> int | float: """simple docstring""" if len(A__ ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(A__ ) or left < -len(A__ ) or right >= len(A__ ) or right < -len(A__ ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] snake_case = (left + right) >> 1 # the middle snake_case = find_max(A__ , A__ , A__ ) # find max in range[left, mid] snake_case = find_max(A__ , mid + 1 , A__ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
137
1
'''simple docstring''' import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase = get_tests_dir('''fixtures/test_sentencepiece.model''') UpperCAmelCase = get_tests_dir('''fixtures/test_sentencepiece_bpe.model''') UpperCAmelCase = '''pt''' if is_torch_available() else '''tf''' @require_sentencepiece @require_tokenizers class lowerCAmelCase ( A , unittest.TestCase ): lowerCAmelCase_ = CamembertTokenizer lowerCAmelCase_ = CamembertTokenizerFast lowerCAmelCase_ = True lowerCAmelCase_ = True def snake_case ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __lowercase =CamembertTokenizer(__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case ( self : List[Any] ): """simple docstring""" __lowercase ='<pad>' __lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) , __lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) , __lowercase ) def snake_case ( self : List[Any] ): """simple docstring""" __lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>NOTUSED' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(__lowercase ) , 1004 ) def snake_case ( self : Tuple ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1005 ) def snake_case ( self : Optional[int] ): """simple docstring""" __lowercase =CamembertTokenizer(__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) __lowercase =CamembertTokenizerFast.from_pretrained(self.tmpdirname ) __lowercase ='I was born in 92000, and this is falsé.' __lowercase =tokenizer.encode(__lowercase ) __lowercase =rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __lowercase =tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __lowercase =rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) __lowercase =tokenizer.convert_ids_to_tokens(__lowercase ) __lowercase =rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def snake_case ( self : Optional[Any] ): """simple docstring""" if not self.test_rust_tokenizer: return __lowercase =self.get_tokenizer() __lowercase =self.get_rust_tokenizer() __lowercase ='I was born in 92000, and this is falsé.' __lowercase =tokenizer.tokenize(__lowercase ) __lowercase =rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __lowercase =tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) __lowercase =rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) __lowercase =self.get_rust_tokenizer() __lowercase =tokenizer.encode(__lowercase ) __lowercase =rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) @slow def snake_case ( self : Dict ): """simple docstring""" __lowercase ={'input_ids': [[5, 54, 7196, 297, 30, 23, 776, 18, 11, 3215, 3705, 8252, 22, 3164, 1181, 2116, 29, 16, 813, 25, 791, 3314, 20, 3446, 38, 27575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9088, 20, 1517, 8, 22804, 18818, 10, 38, 629, 607, 607, 142, 19, 7196, 867, 56, 10326, 24, 2267, 20, 416, 5072, 15612, 233, 734, 7, 2399, 27, 16, 3015, 1649, 7, 24, 20, 4338, 2399, 27, 13, 3400, 14, 13, 6189, 8, 930, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. __lowercase =[ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=__lowercase , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=__lowercase , )
141
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL UpperCAmelCase = logging.get_logger(__name__) def __UpperCamelCase ( lowercase__ : List[Any] ): '''simple docstring''' if isinstance(lowercase__, (list, tuple) ) and isinstance(videos[0], (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowercase__, (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowercase__ ): return [[videos]] raise ValueError(F'''Could not make batched video from {videos}''' ) class lowerCAmelCase ( A ): lowerCAmelCase_ = ["pixel_values"] def __init__( self : Union[str, Any] , __lowercase : bool = True , __lowercase : Dict[str, int] = None , __lowercase : PILImageResampling = PILImageResampling.BILINEAR , __lowercase : bool = True , __lowercase : Dict[str, int] = None , __lowercase : bool = True , __lowercase : Union[int, float] = 1 / 255 , __lowercase : bool = True , __lowercase : Optional[Union[float, List[float]]] = None , __lowercase : Optional[Union[float, List[float]]] = None , **__lowercase : Optional[Any] , ): """simple docstring""" super().__init__(**__lowercase ) __lowercase =size if size is not None else {'shortest_edge': 224} __lowercase =get_size_dict(__lowercase , default_to_square=__lowercase ) __lowercase =crop_size if crop_size is not None else {'height': 224, 'width': 224} __lowercase =get_size_dict(__lowercase , param_name='crop_size' ) __lowercase =do_resize __lowercase =size __lowercase =do_center_crop __lowercase =crop_size __lowercase =resample __lowercase =do_rescale __lowercase =rescale_factor __lowercase =do_normalize __lowercase =image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowercase =image_std if image_std is not None else IMAGENET_STANDARD_STD def snake_case ( self : int , __lowercase : np.ndarray , __lowercase : Dict[str, int] , __lowercase : PILImageResampling = PILImageResampling.BILINEAR , __lowercase : Optional[Union[str, ChannelDimension]] = None , **__lowercase : Optional[Any] , ): """simple docstring""" __lowercase =get_size_dict(__lowercase , default_to_square=__lowercase ) if "shortest_edge" in size: __lowercase =get_resize_output_image_size(__lowercase , size['shortest_edge'] , default_to_square=__lowercase ) elif "height" in size and "width" in size: __lowercase =(size['height'], size['width']) else: raise ValueError(f'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(__lowercase , size=__lowercase , resample=__lowercase , data_format=__lowercase , **__lowercase ) def snake_case ( self : Dict , __lowercase : np.ndarray , __lowercase : Dict[str, int] , __lowercase : Optional[Union[str, ChannelDimension]] = None , **__lowercase : List[Any] , ): """simple docstring""" __lowercase =get_size_dict(__lowercase ) if "height" not in size or "width" not in size: raise ValueError(f'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(__lowercase , size=(size['height'], size['width']) , data_format=__lowercase , **__lowercase ) def snake_case ( self : str , __lowercase : np.ndarray , __lowercase : Union[int, float] , __lowercase : Optional[Union[str, ChannelDimension]] = None , **__lowercase : Any , ): """simple docstring""" return rescale(__lowercase , scale=__lowercase , data_format=__lowercase , **__lowercase ) def snake_case ( self : Dict , __lowercase : np.ndarray , __lowercase : Union[float, List[float]] , __lowercase : Union[float, List[float]] , __lowercase : Optional[Union[str, ChannelDimension]] = None , **__lowercase : Optional[Any] , ): """simple docstring""" return normalize(__lowercase , mean=__lowercase , std=__lowercase , data_format=__lowercase , **__lowercase ) def snake_case ( self : Optional[Any] , __lowercase : ImageInput , __lowercase : bool = None , __lowercase : Dict[str, int] = None , __lowercase : PILImageResampling = None , __lowercase : bool = None , __lowercase : Dict[str, int] = None , __lowercase : bool = None , __lowercase : float = None , __lowercase : bool = None , __lowercase : Optional[Union[float, List[float]]] = None , __lowercase : Optional[Union[float, List[float]]] = None , __lowercase : Optional[ChannelDimension] = ChannelDimension.FIRST , ): """simple docstring""" if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. __lowercase =to_numpy_array(__lowercase ) if do_resize: __lowercase =self.resize(image=__lowercase , size=__lowercase , resample=__lowercase ) if do_center_crop: __lowercase =self.center_crop(__lowercase , size=__lowercase ) if do_rescale: __lowercase =self.rescale(image=__lowercase , scale=__lowercase ) if do_normalize: __lowercase =self.normalize(image=__lowercase , mean=__lowercase , std=__lowercase ) __lowercase =to_channel_dimension_format(__lowercase , __lowercase ) return image def snake_case ( self : Union[str, Any] , __lowercase : ImageInput , __lowercase : bool = None , __lowercase : Dict[str, int] = None , __lowercase : PILImageResampling = None , __lowercase : bool = None , __lowercase : Dict[str, int] = None , __lowercase : bool = None , __lowercase : float = None , __lowercase : bool = None , __lowercase : Optional[Union[float, List[float]]] = None , __lowercase : Optional[Union[float, List[float]]] = None , __lowercase : Optional[Union[str, TensorType]] = None , __lowercase : ChannelDimension = ChannelDimension.FIRST , **__lowercase : Tuple , ): """simple docstring""" __lowercase =do_resize if do_resize is not None else self.do_resize __lowercase =resample if resample is not None else self.resample __lowercase =do_center_crop if do_center_crop is not None else self.do_center_crop __lowercase =do_rescale if do_rescale is not None else self.do_rescale __lowercase =rescale_factor if rescale_factor is not None else self.rescale_factor __lowercase =do_normalize if do_normalize is not None else self.do_normalize __lowercase =image_mean if image_mean is not None else self.image_mean __lowercase =image_std if image_std is not None else self.image_std __lowercase =size if size is not None else self.size __lowercase =get_size_dict(__lowercase , default_to_square=__lowercase ) __lowercase =crop_size if crop_size is not None else self.crop_size __lowercase =get_size_dict(__lowercase , param_name='crop_size' ) if not valid_images(__lowercase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) __lowercase =make_batched(__lowercase ) __lowercase =[ [ self._preprocess_image( image=__lowercase , do_resize=__lowercase , size=__lowercase , resample=__lowercase , do_center_crop=__lowercase , crop_size=__lowercase , do_rescale=__lowercase , rescale_factor=__lowercase , do_normalize=__lowercase , image_mean=__lowercase , image_std=__lowercase , data_format=__lowercase , ) for img in video ] for video in videos ] __lowercase ={'pixel_values': videos} return BatchFeature(data=__lowercase , tensor_type=__lowercase )
141
1
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class UpperCamelCase_ ( unittest.TestCase ): def _lowercase( self ) -> Any: # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. UpperCAmelCase : int = [[1, 2, 4], [1, 2, 3, 4]] UpperCAmelCase : List[str] = DisjunctiveConstraint(snake_case_ ) self.assertTrue(isinstance(dc.token_ids , snake_case_ ) ) with self.assertRaises(snake_case_ ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(snake_case_ ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def _lowercase( self ) -> Optional[Any]: # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). UpperCAmelCase : Union[str, Any] = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(snake_case_ ): DisjunctiveConstraint(snake_case_ ) # fails here def _lowercase( self ) -> str: UpperCAmelCase : List[Any] = [[1, 2, 3], [1, 2, 4]] UpperCAmelCase : Tuple = DisjunctiveConstraint(snake_case_ ) UpperCAmelCase : int = dc.update(1 ) UpperCAmelCase : int = stepped is True and completed is False and reset is False self.assertTrue(snake_case_ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) UpperCAmelCase : int = dc.update(2 ) UpperCAmelCase : int = stepped is True and completed is False and reset is False self.assertTrue(snake_case_ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) UpperCAmelCase : str = dc.update(3 ) UpperCAmelCase : Union[str, Any] = stepped is True and completed is True and reset is False self.assertTrue(snake_case_ ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def _lowercase( self ) -> int: UpperCAmelCase : Tuple = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] UpperCAmelCase : str = DisjunctiveConstraint(snake_case_ ) UpperCAmelCase : Dict = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) UpperCAmelCase : List[str] = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) UpperCAmelCase : Optional[Any] = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) UpperCAmelCase : Union[str, Any] = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() UpperCAmelCase : Union[str, Any] = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) UpperCAmelCase : Optional[Any] = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) UpperCAmelCase : str = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
371
'''simple docstring''' from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer a : List[Any] = logging.get_logger(__name__) a : List[str] = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } a : List[Any] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } a : List[Any] = { """facebook/blenderbot_small-90M""": 5_1_2, } class UpperCamelCase_ ( __magic_name__ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = BlenderbotSmallTokenizer def __init__( self , A=None , A=None , A="<|endoftext|>" , A="<|endoftext|>" , A="<|endoftext|>" , A=False , A=True , **A , ) -> Union[str, Any]: super().__init__( ByteLevelBPETokenizer( vocab=A , merges=A , add_prefix_space=A , trim_offsets=A , ) , bos_token=A , eos_token=A , unk_token=A , **A , ) UpperCAmelCase : Optional[Any] = add_prefix_space def _lowercase( self , A , A=None ) -> Optional[Any]: UpperCAmelCase : Optional[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowercase( self , A , A = None ) -> List[int]: UpperCAmelCase : Any = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
338
0
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, require_torch_gpu, require_torch_or_tf, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): _a = MODEL_FOR_CAUSAL_LM_MAPPING _a = TF_MODEL_FOR_CAUSAL_LM_MAPPING @require_torch def __lowercase ( self : Optional[Any] ): lowerCAmelCase = pipeline(task="""text-generation""" , model="""sshleifer/tiny-ctrl""" , framework="""pt""" ) # Using `do_sample=False` to force deterministic output lowerCAmelCase = text_generator("""This is a test""" , do_sample=lowerCAmelCase ) self.assertEqual( lowerCAmelCase , [ { """generated_text""": ( """This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.""" """ oscope. FiliFili@@""" ) } ] , ) lowerCAmelCase = text_generator(["""This is a test""", """This is a second test"""] ) self.assertEqual( lowerCAmelCase , [ [ { """generated_text""": ( """This is a test ☃ ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy oscope.""" """ oscope. FiliFili@@""" ) } ], [ { """generated_text""": ( """This is a second test ☃ segmental segmental segmental 议议eski eski flutter flutter Lacy""" """ oscope. oscope. FiliFili@@""" ) } ], ] , ) lowerCAmelCase = text_generator("""This is a test""" , do_sample=lowerCAmelCase , num_return_sequences=2 , return_tensors=lowerCAmelCase ) self.assertEqual( lowerCAmelCase , [ {"""generated_token_ids""": ANY(lowerCAmelCase )}, {"""generated_token_ids""": ANY(lowerCAmelCase )}, ] , ) lowerCAmelCase = text_generator.model.config.eos_token_id lowerCAmelCase = """<pad>""" lowerCAmelCase = text_generator( ["""This is a test""", """This is a second test"""] , do_sample=lowerCAmelCase , num_return_sequences=2 , batch_size=2 , return_tensors=lowerCAmelCase , ) self.assertEqual( lowerCAmelCase , [ [ {"""generated_token_ids""": ANY(lowerCAmelCase )}, {"""generated_token_ids""": ANY(lowerCAmelCase )}, ], [ {"""generated_token_ids""": ANY(lowerCAmelCase )}, {"""generated_token_ids""": ANY(lowerCAmelCase )}, ], ] , ) @require_tf def __lowercase ( self : List[str] ): lowerCAmelCase = pipeline(task="""text-generation""" , model="""sshleifer/tiny-ctrl""" , framework="""tf""" ) # Using `do_sample=False` to force deterministic output lowerCAmelCase = text_generator("""This is a test""" , do_sample=lowerCAmelCase ) self.assertEqual( lowerCAmelCase , [ { """generated_text""": ( """This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵""" """ please,""" ) } ] , ) lowerCAmelCase = text_generator(["""This is a test""", """This is a second test"""] , do_sample=lowerCAmelCase ) self.assertEqual( lowerCAmelCase , [ [ { """generated_text""": ( """This is a test FeyFeyFey(Croatis.), s.), Cannes Cannes Cannes 閲閲Cannes Cannes Cannes 攵""" """ please,""" ) } ], [ { """generated_text""": ( """This is a second test Chieftain Chieftain prefecture prefecture prefecture Cannes Cannes""" """ Cannes 閲閲Cannes Cannes Cannes 攵 please,""" ) } ], ] , ) def __lowercase ( self : List[Any] , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : Optional[Any] , lowerCAmelCase : str ): lowerCAmelCase = TextGenerationPipeline(model=lowerCAmelCase , tokenizer=lowerCAmelCase ) return text_generator, ["This is a test", "Another test"] def __lowercase ( self : Union[str, Any] ): lowerCAmelCase = """Hello I believe in""" lowerCAmelCase = pipeline("""text-generation""" , model="""hf-internal-testing/tiny-random-gpt2""" ) lowerCAmelCase = text_generator(lowerCAmelCase ) self.assertEqual( lowerCAmelCase , [{"""generated_text""": """Hello I believe in fe fe fe fe fe fe fe fe fe fe fe fe"""}] , ) lowerCAmelCase = text_generator(lowerCAmelCase , stop_sequence=""" fe""" ) self.assertEqual(lowerCAmelCase , [{"""generated_text""": """Hello I believe in fe"""}] ) def __lowercase ( self : Union[str, Any] , lowerCAmelCase : List[Any] , lowerCAmelCase : Optional[Any] ): lowerCAmelCase = text_generator.model lowerCAmelCase = text_generator.tokenizer lowerCAmelCase = text_generator("""This is a test""" ) self.assertEqual(lowerCAmelCase , [{"""generated_text""": ANY(lowerCAmelCase )}] ) self.assertTrue(outputs[0]["""generated_text"""].startswith("""This is a test""" ) ) lowerCAmelCase = text_generator("""This is a test""" , return_full_text=lowerCAmelCase ) self.assertEqual(lowerCAmelCase , [{"""generated_text""": ANY(lowerCAmelCase )}] ) self.assertNotIn("""This is a test""" , outputs[0]["""generated_text"""] ) lowerCAmelCase = pipeline(task="""text-generation""" , model=lowerCAmelCase , tokenizer=lowerCAmelCase , return_full_text=lowerCAmelCase ) lowerCAmelCase = text_generator("""This is a test""" ) self.assertEqual(lowerCAmelCase , [{"""generated_text""": ANY(lowerCAmelCase )}] ) self.assertNotIn("""This is a test""" , outputs[0]["""generated_text"""] ) lowerCAmelCase = text_generator("""This is a test""" , return_full_text=lowerCAmelCase ) self.assertEqual(lowerCAmelCase , [{"""generated_text""": ANY(lowerCAmelCase )}] ) self.assertTrue(outputs[0]["""generated_text"""].startswith("""This is a test""" ) ) lowerCAmelCase = text_generator(["""This is great !""", """Something else"""] , num_return_sequences=2 , do_sample=lowerCAmelCase ) self.assertEqual( lowerCAmelCase , [ [{"""generated_text""": ANY(lowerCAmelCase )}, {"""generated_text""": ANY(lowerCAmelCase )}], [{"""generated_text""": ANY(lowerCAmelCase )}, {"""generated_text""": ANY(lowerCAmelCase )}], ] , ) if text_generator.tokenizer.pad_token is not None: lowerCAmelCase = text_generator( ["""This is great !""", """Something else"""] , num_return_sequences=2 , batch_size=2 , do_sample=lowerCAmelCase ) self.assertEqual( lowerCAmelCase , [ [{"""generated_text""": ANY(lowerCAmelCase )}, {"""generated_text""": ANY(lowerCAmelCase )}], [{"""generated_text""": ANY(lowerCAmelCase )}, {"""generated_text""": ANY(lowerCAmelCase )}], ] , ) with self.assertRaises(lowerCAmelCase ): lowerCAmelCase = text_generator("""test""" , return_full_text=lowerCAmelCase , return_text=lowerCAmelCase ) with self.assertRaises(lowerCAmelCase ): lowerCAmelCase = text_generator("""test""" , return_full_text=lowerCAmelCase , return_tensors=lowerCAmelCase ) with self.assertRaises(lowerCAmelCase ): lowerCAmelCase = text_generator("""test""" , return_text=lowerCAmelCase , return_tensors=lowerCAmelCase ) # Empty prompt is slighly special # it requires BOS token to exist. # Special case for Pegasus which will always append EOS so will # work even without BOS. if ( text_generator.tokenizer.bos_token_id is not None or "Pegasus" in tokenizer.__class__.__name__ or "Git" in model.__class__.__name__ ): lowerCAmelCase = text_generator("""""" ) self.assertEqual(lowerCAmelCase , [{"""generated_text""": ANY(lowerCAmelCase )}] ) else: with self.assertRaises((ValueError, AssertionError) ): lowerCAmelCase = text_generator("""""" ) if text_generator.framework == "tf": # TF generation does not support max_new_tokens, and it's impossible # to control long generation with only max_length without # fancy calculation, dismissing tests for now. return # We don't care about infinite range models. # They already work. # Skip this test for XGLM, since it uses sinusoidal positional embeddings which are resized on-the-fly. lowerCAmelCase = ["""RwkvForCausalLM""", """XGLMForCausalLM""", """GPTNeoXForCausalLM"""] if ( tokenizer.model_max_length < 1_0000 and text_generator.model.__class__.__name__ not in EXTRA_MODELS_CAN_HANDLE_LONG_INPUTS ): # Handling of large generations with self.assertRaises((RuntimeError, IndexError, ValueError, AssertionError) ): text_generator("""This is a test""" * 500 , max_new_tokens=20 ) lowerCAmelCase = text_generator("""This is a test""" * 500 , handle_long_generation="""hole""" , max_new_tokens=20 ) # Hole strategy cannot work with self.assertRaises(lowerCAmelCase ): text_generator( """This is a test""" * 500 , handle_long_generation="""hole""" , max_new_tokens=tokenizer.model_max_length + 10 , ) @require_torch @require_accelerate @require_torch_gpu def __lowercase ( self : Tuple ): import torch # Classic `model_kwargs` lowerCAmelCase = pipeline( model="""hf-internal-testing/tiny-random-bloom""" , model_kwargs={"""device_map""": """auto""", """torch_dtype""": torch.bfloataa} , ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) lowerCAmelCase = pipe("""This is a test""" ) self.assertEqual( lowerCAmelCase , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) # Upgraded those two to real pipeline arguments (they just get sent for the model as they're unlikely to mean anything else.) lowerCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" , torch_dtype=torch.bfloataa ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.bfloataa ) lowerCAmelCase = pipe("""This is a test""" ) self.assertEqual( lowerCAmelCase , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) # torch_dtype will be automatically set to float32 if not provided - check: https://github.com/huggingface/transformers/pull/20602 lowerCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" ) self.assertEqual(pipe.model.device , torch.device(0 ) ) self.assertEqual(pipe.model.lm_head.weight.dtype , torch.floataa ) lowerCAmelCase = pipe("""This is a test""" ) self.assertEqual( lowerCAmelCase , [ { """generated_text""": ( """This is a test test test test test test test test test test test test test test test test""" """ test""" ) } ] , ) @require_torch @require_torch_gpu def __lowercase ( self : List[Any] ): import torch lowerCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device=0 , torch_dtype=torch.floataa ) pipe("""This is a test""" ) @require_torch @require_accelerate @require_torch_gpu def __lowercase ( self : List[str] ): import torch lowerCAmelCase = pipeline(model="""hf-internal-testing/tiny-random-bloom""" , device_map="""auto""" , torch_dtype=torch.floataa ) pipe("""This is a test""" , do_sample=lowerCAmelCase , top_p=0.5 ) def __lowercase ( self : Dict ): lowerCAmelCase = """Hello world""" lowerCAmelCase = pipeline("""text-generation""" , model="""hf-internal-testing/tiny-random-gpt2""" ) if text_generator.model.framework == "tf": lowerCAmelCase = logging.get_logger("""transformers.generation.tf_utils""" ) else: lowerCAmelCase = logging.get_logger("""transformers.generation.utils""" ) lowerCAmelCase = """Both `max_new_tokens`""" # The beggining of the message to be checked in this test # Both are set by the user -> log warning with CaptureLogger(lowerCAmelCase ) as cl: lowerCAmelCase = text_generator(lowerCAmelCase , max_length=10 , max_new_tokens=1 ) self.assertIn(lowerCAmelCase , cl.out ) # The user only sets one -> no warning with CaptureLogger(lowerCAmelCase ) as cl: lowerCAmelCase = text_generator(lowerCAmelCase , max_new_tokens=1 ) self.assertNotIn(lowerCAmelCase , cl.out ) with CaptureLogger(lowerCAmelCase ) as cl: lowerCAmelCase = text_generator(lowerCAmelCase , max_length=10 ) self.assertNotIn(lowerCAmelCase , cl.out )
155
"""simple docstring""" from __future__ import annotations def lowercase (snake_case__ : list , snake_case__ : int , snake_case__ : int , snake_case__ : int ) -> list: '''simple docstring''' lowerCAmelCase = [] lowerCAmelCase , lowerCAmelCase = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) lowerCAmelCase = result + left + right return input_list def lowercase (snake_case__ : list ) -> list: '''simple docstring''' if len(snake_case__ ) <= 1: return input_list lowerCAmelCase = list(snake_case__ ) # iteration for two-way merging lowerCAmelCase = 2 while p <= len(snake_case__ ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(snake_case__ ) , snake_case__ ): lowerCAmelCase = i lowerCAmelCase = i + p - 1 lowerCAmelCase = (low + high + 1) // 2 lowerCAmelCase = merge(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) # final merge of last two parts if p * 2 >= len(snake_case__ ): lowerCAmelCase = i lowerCAmelCase = merge(snake_case__ , 0 , snake_case__ , len(snake_case__ ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": a = input('Enter numbers separated by a comma:\n').strip() if user_input == "": a = [] else: a = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
155
1
import datasets from .evaluate import evaluate __A = "\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n" __A = "\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n" __A = "\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair as given in the references (see below)\n - 'prediction_text': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair (see above),\n - 'answers': a Dict in the SQuAD dataset format\n {\n 'text': list of possible texts for the answer, as a list of strings\n 'answer_start': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n 'exact_match': Exact match (the normalized answer exactly match the gold answer)\n 'f1': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{'prediction_text': '1976', 'id': '56e10a3be3433e1400422b22'}]\n >>> references = [{'answers': {'answer_start': [97], 'text': ['1976']}, 'id': '56e10a3be3433e1400422b22'}]\n >>> squad_metric = datasets.load_metric(\"squad\")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 100.0, 'f1': 100.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _SCREAMING_SNAKE_CASE ( datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE_ (self : int) ->Optional[Any]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": {"id": datasets.Value("string"), "prediction_text": datasets.Value("string")}, "references": { "id": datasets.Value("string"), "answers": datasets.features.Sequence( { "text": datasets.Value("string"), "answer_start": datasets.Value("int32"), }), }, }) , codebase_urls=["https://rajpurkar.github.io/SQuAD-explorer/"] , reference_urls=["https://rajpurkar.github.io/SQuAD-explorer/"] , ) def SCREAMING_SNAKE_CASE_ (self : Any , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[int]) ->Optional[int]: '''simple docstring''' lowerCamelCase__: Optional[Any] ={prediction["id"]: prediction["prediction_text"] for prediction in predictions} lowerCamelCase__: str =[ { "paragraphs": [ { "qas": [ { "answers": [{"text": answer_text} for answer_text in ref["answers"]["text"]], "id": ref["id"], } for ref in references ] } ] } ] lowerCamelCase__: Any =evaluate(dataset=lowercase_ , predictions=lowercase_) return score
364
def lowerCAmelCase_ ( __a ) -> str: """simple docstring""" if isinstance(__a , __a ): raise TypeError("'float' object cannot be interpreted as an integer" ) if isinstance(__a , __a ): raise TypeError("'str' object cannot be interpreted as an integer" ) if num == 0: return "0b0" lowerCamelCase__: Optional[int] =False if num < 0: lowerCamelCase__: Optional[Any] =True lowerCamelCase__: List[Any] =-num lowerCamelCase__: list[int] =[] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(__a ) for e in binary ) return "0b" + "".join(str(__a ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
273
0
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class _SCREAMING_SNAKE_CASE( unittest.TestCase ): def _UpperCamelCase ( self ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE :Any = tempfile.mkdtemp() __SCREAMING_SNAKE_CASE :int = BlipImageProcessor() __SCREAMING_SNAKE_CASE :int = BertTokenizer.from_pretrained('''hf-internal-testing/tiny-random-BertModel''' ) __SCREAMING_SNAKE_CASE :Any = BlipProcessor(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) processor.save_pretrained(self.tmpdirname ) def _UpperCamelCase ( self ,**SCREAMING_SNAKE_CASE__ ) -> Tuple: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname ,**SCREAMING_SNAKE_CASE__ ).tokenizer def _UpperCamelCase ( self ,**SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname ,**SCREAMING_SNAKE_CASE__ ).image_processor def _UpperCamelCase ( self ) -> Dict: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = [np.random.randint(2_55 ,size=(3, 30, 4_00) ,dtype=np.uinta )] __SCREAMING_SNAKE_CASE :Optional[int] = [Image.fromarray(np.moveaxis(SCREAMING_SNAKE_CASE__ ,0 ,-1 ) ) for x in image_inputs] return image_inputs def _UpperCamelCase ( self ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE :Any = BlipProcessor(tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __SCREAMING_SNAKE_CASE :List[str] = self.get_tokenizer(bos_token='''(BOS)''' ,eos_token='''(EOS)''' ) __SCREAMING_SNAKE_CASE :int = self.get_image_processor(do_normalize=SCREAMING_SNAKE_CASE__ ,padding_value=1.0 ) __SCREAMING_SNAKE_CASE :List[str] = BlipProcessor.from_pretrained( self.tmpdirname ,bos_token='''(BOS)''' ,eos_token='''(EOS)''' ,do_normalize=SCREAMING_SNAKE_CASE__ ,padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer ,SCREAMING_SNAKE_CASE__ ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = self.get_image_processor() __SCREAMING_SNAKE_CASE :str = self.get_tokenizer() __SCREAMING_SNAKE_CASE :Tuple = BlipProcessor(tokenizer=SCREAMING_SNAKE_CASE__ ,image_processor=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = self.prepare_image_inputs() __SCREAMING_SNAKE_CASE :Optional[int] = image_processor(SCREAMING_SNAKE_CASE__ ,return_tensors='''np''' ) __SCREAMING_SNAKE_CASE :str = processor(images=SCREAMING_SNAKE_CASE__ ,return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1E-2 ) def _UpperCamelCase ( self ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE :str = self.get_image_processor() __SCREAMING_SNAKE_CASE :str = self.get_tokenizer() __SCREAMING_SNAKE_CASE :str = BlipProcessor(tokenizer=SCREAMING_SNAKE_CASE__ ,image_processor=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[str] = '''lower newer''' __SCREAMING_SNAKE_CASE :Union[str, Any] = processor(text=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = tokenizer(SCREAMING_SNAKE_CASE__ ,return_token_type_ids=SCREAMING_SNAKE_CASE__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :int = self.get_image_processor() __SCREAMING_SNAKE_CASE :Any = self.get_tokenizer() __SCREAMING_SNAKE_CASE :Optional[Any] = BlipProcessor(tokenizer=SCREAMING_SNAKE_CASE__ ,image_processor=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[Any] = '''lower newer''' __SCREAMING_SNAKE_CASE :str = self.prepare_image_inputs() __SCREAMING_SNAKE_CASE :List[str] = processor(text=SCREAMING_SNAKE_CASE__ ,images=SCREAMING_SNAKE_CASE__ ) self.assertListEqual(list(inputs.keys() ) ,['''pixel_values''', '''input_ids''', '''attention_mask'''] ) # test if it raises when no input is passed with pytest.raises(SCREAMING_SNAKE_CASE__ ): processor() def _UpperCamelCase ( self ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[str] = self.get_image_processor() __SCREAMING_SNAKE_CASE :str = self.get_tokenizer() __SCREAMING_SNAKE_CASE :List[str] = BlipProcessor(tokenizer=SCREAMING_SNAKE_CASE__ ,image_processor=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __SCREAMING_SNAKE_CASE :Union[str, Any] = processor.batch_decode(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = tokenizer.batch_decode(SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = self.get_image_processor() __SCREAMING_SNAKE_CASE :int = self.get_tokenizer() __SCREAMING_SNAKE_CASE :List[str] = BlipProcessor(tokenizer=SCREAMING_SNAKE_CASE__ ,image_processor=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = '''lower newer''' __SCREAMING_SNAKE_CASE :Optional[int] = self.prepare_image_inputs() __SCREAMING_SNAKE_CASE :Tuple = processor(text=SCREAMING_SNAKE_CASE__ ,images=SCREAMING_SNAKE_CASE__ ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) ,['''pixel_values''', '''input_ids''', '''attention_mask'''] )
191
"""simple docstring""" import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _SCREAMING_SNAKE_CASE( A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[Any] = ConsistencyModelPipeline SCREAMING_SNAKE_CASE_ : Any = UNCONDITIONAL_IMAGE_GENERATION_PARAMS SCREAMING_SNAKE_CASE_ : Dict = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt SCREAMING_SNAKE_CASE_ : Optional[Any] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''output_type''', '''return_dict''', '''callback''', '''callback_steps''', ] ) @property def _UpperCamelCase ( self ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' ,subfolder='''test_unet''' ,) return unet @property def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' ,subfolder='''test_unet_class_cond''' ,) return unet def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__=False ) -> Union[str, Any]: """simple docstring""" if class_cond: __SCREAMING_SNAKE_CASE :str = self.dummy_cond_unet else: __SCREAMING_SNAKE_CASE :Optional[Any] = self.dummy_uncond_unet # Default to CM multistep sampler __SCREAMING_SNAKE_CASE :List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 ,sigma_min=0.0_0_2 ,sigma_max=8_0.0 ,) __SCREAMING_SNAKE_CASE :List[str] = { '''unet''': unet, '''scheduler''': scheduler, } return components def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__=0 ) -> Dict: """simple docstring""" if str(SCREAMING_SNAKE_CASE__ ).startswith('''mps''' ): __SCREAMING_SNAKE_CASE :Tuple = torch.manual_seed(SCREAMING_SNAKE_CASE__ ) else: __SCREAMING_SNAKE_CASE :Optional[Any] = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = { '''batch_size''': 1, '''num_inference_steps''': None, '''timesteps''': [22, 0], '''generator''': generator, '''output_type''': '''np''', } return inputs def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :List[str] = self.get_dummy_components() __SCREAMING_SNAKE_CASE :Optional[Any] = ConsistencyModelPipeline(**SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Tuple = self.get_dummy_inputs(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE :List[str] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :Any = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE :Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :List[Any] = self.get_dummy_components(class_cond=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = ConsistencyModelPipeline(**SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = self.get_dummy_inputs(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = 0 __SCREAMING_SNAKE_CASE :Optional[int] = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE :Dict = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :List[Any] = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :Tuple = self.get_dummy_components() __SCREAMING_SNAKE_CASE :Any = ConsistencyModelPipeline(**SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[str] = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = self.get_dummy_inputs(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = 1 __SCREAMING_SNAKE_CASE :List[str] = None __SCREAMING_SNAKE_CASE :List[str] = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE :List[str] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :int = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :str = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :Any = self.get_dummy_components(class_cond=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = ConsistencyModelPipeline(**SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = self.get_dummy_inputs(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = 1 __SCREAMING_SNAKE_CASE :Optional[Any] = None __SCREAMING_SNAKE_CASE :List[Any] = 0 __SCREAMING_SNAKE_CASE :Any = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE :int = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :Optional[Any] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE( unittest.TestCase ): def _UpperCamelCase ( self ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__=0 ,SCREAMING_SNAKE_CASE__=False ,SCREAMING_SNAKE_CASE__="cpu" ,SCREAMING_SNAKE_CASE__=torch.floataa ,SCREAMING_SNAKE_CASE__=(1, 3, 64, 64) ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = torch.manual_seed(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Tuple = { '''num_inference_steps''': None, '''timesteps''': [22, 0], '''class_labels''': 0, '''generator''': generator, '''output_type''': '''np''', } if get_fixed_latents: __SCREAMING_SNAKE_CASE :int = self.get_fixed_latents(seed=SCREAMING_SNAKE_CASE__ ,device=SCREAMING_SNAKE_CASE__ ,dtype=SCREAMING_SNAKE_CASE__ ,shape=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = latents return inputs def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__=0 ,SCREAMING_SNAKE_CASE__="cpu" ,SCREAMING_SNAKE_CASE__=torch.floataa ,SCREAMING_SNAKE_CASE__=(1, 3, 64, 64) ) -> int: """simple docstring""" if type(SCREAMING_SNAKE_CASE__ ) == str: __SCREAMING_SNAKE_CASE :int = torch.device(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = randn_tensor(SCREAMING_SNAKE_CASE__ ,generator=SCREAMING_SNAKE_CASE__ ,device=SCREAMING_SNAKE_CASE__ ,dtype=SCREAMING_SNAKE_CASE__ ) return latents def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' ,subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE :List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 ,sigma_min=0.0_0_2 ,sigma_max=8_0.0 ,) __SCREAMING_SNAKE_CASE :Dict = ConsistencyModelPipeline(unet=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ) pipe.to(torch_device=SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = self.get_inputs() __SCREAMING_SNAKE_CASE :List[str] = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE :Union[str, Any] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :Dict = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' ,subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE :List[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 ,sigma_min=0.0_0_2 ,sigma_max=8_0.0 ,) __SCREAMING_SNAKE_CASE :Any = ConsistencyModelPipeline(unet=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ) pipe.to(torch_device=SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = self.get_inputs() __SCREAMING_SNAKE_CASE :int = 1 __SCREAMING_SNAKE_CASE :int = None __SCREAMING_SNAKE_CASE :Union[str, Any] = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE :str = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :List[str] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' ,subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE :Any = CMStochasticIterativeScheduler( num_train_timesteps=40 ,sigma_min=0.0_0_2 ,sigma_max=8_0.0 ,) __SCREAMING_SNAKE_CASE :Any = ConsistencyModelPipeline(unet=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ) pipe.to(torch_device=SCREAMING_SNAKE_CASE__ ,torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :str = self.get_inputs(get_fixed_latents=SCREAMING_SNAKE_CASE__ ,device=SCREAMING_SNAKE_CASE__ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=SCREAMING_SNAKE_CASE__ ,enable_math=SCREAMING_SNAKE_CASE__ ,enable_mem_efficient=SCREAMING_SNAKE_CASE__ ): __SCREAMING_SNAKE_CASE :Optional[Any] = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE :List[str] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :List[Any] = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :str = UNetaDModel.from_pretrained('''diffusers/consistency_models''' ,subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE :Dict = CMStochasticIterativeScheduler( num_train_timesteps=40 ,sigma_min=0.0_0_2 ,sigma_max=8_0.0 ,) __SCREAMING_SNAKE_CASE :int = ConsistencyModelPipeline(unet=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ) pipe.to(torch_device=SCREAMING_SNAKE_CASE__ ,torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = self.get_inputs(get_fixed_latents=SCREAMING_SNAKE_CASE__ ,device=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = 1 __SCREAMING_SNAKE_CASE :int = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=SCREAMING_SNAKE_CASE__ ,enable_math=SCREAMING_SNAKE_CASE__ ,enable_mem_efficient=SCREAMING_SNAKE_CASE__ ): __SCREAMING_SNAKE_CASE :str = pipe(**SCREAMING_SNAKE_CASE__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE :str = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :Optional[int] = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
191
1
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __lowercase = """\ @inproceedings{kakwani2020indicnlpsuite, title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}}, author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar}, year={2020}, booktitle={Findings of EMNLP}, } """ __lowercase = """\ IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te. """ __lowercase = """ Compute IndicGLUE evaluation metric associated to each IndicGLUE dataset. Args: predictions: list of predictions to score (as int64), except for 'cvit-mkb-clsr' where each prediction is a vector (of float32). references: list of ground truth labels corresponding to the predictions (as int64), except for 'cvit-mkb-clsr' where each reference is a vector (of float32). Returns: depending on the IndicGLUE subset, one or several of: \"accuracy\": Accuracy \"f1\": F1 score \"precision\": Precision@10 Examples: >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wnli') # 'wnli' or any of [\"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wiki-ner') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'cvit-mkb-clsr') >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'precision@10': 1.0} """ def lowercase ( A_ , A_ )-> Union[str, Any]: '''simple docstring''' return float((preds == labels).mean() ) def lowercase ( A_ , A_ )-> str: '''simple docstring''' a : int = simple_accuracy(A_ , A_ ) a : Optional[int] = float(fa_score(y_true=A_ , y_pred=A_ ) ) return { "accuracy": acc, "f1": fa, } def lowercase ( A_ , A_ )-> int: '''simple docstring''' a : Union[str, Any] = np.array(A_ ) a : List[str] = np.array(A_ ) a : Tuple = en_sentvecs.shape[0] # mean centering a : Tuple = en_sentvecs - np.mean(A_ , axis=0 ) a : Dict = in_sentvecs - np.mean(A_ , axis=0 ) a : Optional[Any] = cdist(A_ , A_ , "cosine" ) a : Optional[Any] = np.array(range(A_ ) ) a : Optional[Any] = sim.argsort(axis=1 )[:, :10] a : Any = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class _A ( datasets.Metric ): """simple docstring""" def __snake_case ( self : Union[str, Any]): if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def __snake_case ( self : Any , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Dict): if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(__UpperCAmelCase , __UpperCAmelCase)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(__UpperCAmelCase , __UpperCAmelCase) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(__UpperCAmelCase , __UpperCAmelCase)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
362
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig __lowercase = logging.get_logger(__name__) # General docstring __lowercase = """MobileNetV1Config""" # Base docstring __lowercase = """google/mobilenet_v1_1.0_224""" __lowercase = [1, 1024, 7, 7] # Image classification docstring __lowercase = """google/mobilenet_v1_1.0_224""" __lowercase = """tabby, tabby cat""" __lowercase = [ """google/mobilenet_v1_1.0_224""", """google/mobilenet_v1_0.75_192""", # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def lowercase ( A_ , A_ , A_=None )-> int: '''simple docstring''' a : Union[str, Any] = {} if isinstance(A_ , A_ ): a : Tuple = model.mobilenet_va else: a : Optional[int] = model a : Any = "MobilenetV1/Conv2d_0/" a : List[Any] = backbone.conv_stem.convolution.weight a : Tuple = backbone.conv_stem.normalization.bias a : int = backbone.conv_stem.normalization.weight a : Optional[int] = backbone.conv_stem.normalization.running_mean a : Dict = backbone.conv_stem.normalization.running_var for i in range(13 ): a : Dict = i + 1 a : str = i * 2 a : str = backbone.layer[pt_index] a : Tuple = F'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' a : Union[str, Any] = pointer.convolution.weight a : Optional[int] = pointer.normalization.bias a : Any = pointer.normalization.weight a : Optional[int] = pointer.normalization.running_mean a : Dict = pointer.normalization.running_var a : Dict = backbone.layer[pt_index + 1] a : List[str] = F'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' a : Dict = pointer.convolution.weight a : Any = pointer.normalization.bias a : List[Any] = pointer.normalization.weight a : Tuple = pointer.normalization.running_mean a : List[str] = pointer.normalization.running_var if isinstance(A_ , A_ ): a : Dict = "MobilenetV1/Logits/Conv2d_1c_1x1/" a : List[Any] = model.classifier.weight a : Optional[int] = model.classifier.bias return tf_to_pt_map def lowercase ( A_ , A_ , A_ )-> Optional[int]: '''simple docstring''' try: import numpy as np import tensorflow as tf except ImportError: logger.error( "Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see " "https://www.tensorflow.org/install/ for installation instructions." ) raise # Load weights from TF model a : List[str] = tf.train.list_variables(A_ ) a : Dict = {} for name, shape in init_vars: logger.info(F'''Loading TF weight {name} with shape {shape}''' ) a : List[str] = tf.train.load_variable(A_ , A_ ) a : int = array # Build TF to PyTorch weights loading map a : int = _build_tf_to_pytorch_map(A_ , A_ , A_ ) for name, pointer in tf_to_pt_map.items(): logger.info(F'''Importing {name}''' ) if name not in tf_weights: logger.info(F'''{name} not in tf pre-trained weights, skipping''' ) continue a : Tuple = tf_weights[name] if "depthwise_weights" in name: logger.info("Transposing depthwise" ) a : Optional[Any] = np.transpose(A_ , (2, 3, 0, 1) ) elif "weights" in name: logger.info("Transposing" ) if len(pointer.shape ) == 2: # copying into linear layer a : List[str] = array.squeeze().transpose() else: a : List[str] = np.transpose(A_ , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''' ) logger.info(F'''Initialize PyTorch weight {name} {array.shape}''' ) a : Any = torch.from_numpy(A_ ) tf_weights.pop(A_ , A_ ) tf_weights.pop(name + "/RMSProp" , A_ ) tf_weights.pop(name + "/RMSProp_1" , A_ ) tf_weights.pop(name + "/ExponentialMovingAverage" , A_ ) logger.info(F'''Weights not copied to PyTorch model: {", ".join(tf_weights.keys() )}''' ) return model def lowercase ( A_ , A_ )-> torch.Tensor: '''simple docstring''' a , a : str = features.shape[-2:] a , a : Tuple = conv_layer.stride a , a : Union[str, Any] = conv_layer.kernel_size if in_height % stride_height == 0: a : Union[str, Any] = max(kernel_height - stride_height , 0 ) else: a : Tuple = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: a : Optional[int] = max(kernel_width - stride_width , 0 ) else: a : int = max(kernel_width - (in_width % stride_width) , 0 ) a : List[Any] = pad_along_width // 2 a : List[str] = pad_along_width - pad_left a : str = pad_along_height // 2 a : Any = pad_along_height - pad_top a : Tuple = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(A_ , A_ , "constant" , 0.0 ) class _A ( nn.Module ): """simple docstring""" def __init__( self : Dict , __UpperCAmelCase : MobileNetVaConfig , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : Optional[int] = 1 , __UpperCAmelCase : Optional[int] = 1 , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[bool] = True , __UpperCAmelCase : Optional[bool or str] = True , ): super().__init__() a : Optional[Any] = config if in_channels % groups != 0: raise ValueError(f'''Input channels ({in_channels}) are not divisible by {groups} groups.''') if out_channels % groups != 0: raise ValueError(f'''Output channels ({out_channels}) are not divisible by {groups} groups.''') a : Dict = 0 if config.tf_padding else int((kernel_size - 1) / 2) a : List[str] = nn.Convad( in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , kernel_size=__UpperCAmelCase , stride=__UpperCAmelCase , padding=__UpperCAmelCase , groups=__UpperCAmelCase , bias=__UpperCAmelCase , padding_mode="zeros" , ) if use_normalization: a : str = nn.BatchNormad( num_features=__UpperCAmelCase , eps=config.layer_norm_eps , momentum=0.9_997 , affine=__UpperCAmelCase , track_running_stats=__UpperCAmelCase , ) else: a : List[str] = None if use_activation: if isinstance(__UpperCAmelCase , __UpperCAmelCase): a : Optional[int] = ACTaFN[use_activation] elif isinstance(config.hidden_act , __UpperCAmelCase): a : Optional[Any] = ACTaFN[config.hidden_act] else: a : str = config.hidden_act else: a : Dict = None def __snake_case ( self : Tuple , __UpperCAmelCase : torch.Tensor): if self.config.tf_padding: a : Tuple = apply_tf_padding(__UpperCAmelCase , self.convolution) a : Dict = self.convolution(__UpperCAmelCase) if self.normalization is not None: a : List[str] = self.normalization(__UpperCAmelCase) if self.activation is not None: a : Optional[Any] = self.activation(__UpperCAmelCase) return features class _A ( _a ): """simple docstring""" UpperCAmelCase : Union[str, Any] = MobileNetVaConfig UpperCAmelCase : str = load_tf_weights_in_mobilenet_va UpperCAmelCase : List[str] = """mobilenet_v1""" UpperCAmelCase : Dict = """pixel_values""" UpperCAmelCase : str = False def __snake_case ( self : str , __UpperCAmelCase : Union[nn.Linear, nn.Convad]): if isinstance(__UpperCAmelCase , (nn.Linear, nn.Convad)): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range) if module.bias is not None: module.bias.data.zero_() elif isinstance(__UpperCAmelCase , nn.BatchNormad): module.bias.data.zero_() module.weight.data.fill_(1.0) __lowercase = R""" This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. """ __lowercase = R""" Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. """ @add_start_docstrings( """The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.""" ,_a ,) class _A ( _a ): """simple docstring""" def __init__( self : Optional[int] , __UpperCAmelCase : MobileNetVaConfig , __UpperCAmelCase : bool = True): super().__init__(__UpperCAmelCase) a : List[str] = config a : List[Any] = 32 a : Union[str, Any] = max(int(depth * config.depth_multiplier) , config.min_depth) a : List[str] = MobileNetVaConvLayer( __UpperCAmelCase , in_channels=config.num_channels , out_channels=__UpperCAmelCase , kernel_size=3 , stride=2 , ) a : int = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] a : List[Any] = nn.ModuleList() for i in range(13): a : Optional[Any] = out_channels if strides[i] == 2 or i == 0: depth *= 2 a : List[str] = max(int(depth * config.depth_multiplier) , config.min_depth) self.layer.append( MobileNetVaConvLayer( __UpperCAmelCase , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , kernel_size=3 , stride=strides[i] , groups=__UpperCAmelCase , )) self.layer.append( MobileNetVaConvLayer( __UpperCAmelCase , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , kernel_size=1 , )) a : Optional[Any] = nn.AdaptiveAvgPoolad((1, 1)) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def __snake_case ( self : Optional[Any] , __UpperCAmelCase : Any): raise NotImplementedError @add_start_docstrings_to_model_forward(__UpperCAmelCase) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=__UpperCAmelCase , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def __snake_case ( self : List[Any] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ): a : str = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) a : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values") a : Optional[int] = self.conv_stem(__UpperCAmelCase) a : Tuple = () if output_hidden_states else None for i, layer_module in enumerate(self.layer): a : Union[str, Any] = layer_module(__UpperCAmelCase) if output_hidden_states: a : str = all_hidden_states + (hidden_states,) a : Dict = hidden_states if self.pooler is not None: a : Optional[int] = torch.flatten(self.pooler(__UpperCAmelCase) , start_dim=1) else: a : List[str] = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=__UpperCAmelCase , pooler_output=__UpperCAmelCase , hidden_states=__UpperCAmelCase , ) @add_start_docstrings( """ MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ ,_a ,) class _A ( _a ): """simple docstring""" def __init__( self : int , __UpperCAmelCase : MobileNetVaConfig): super().__init__(__UpperCAmelCase) a : Dict = config.num_labels a : Union[str, Any] = MobileNetVaModel(__UpperCAmelCase) a : List[Any] = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head a : Union[str, Any] = nn.Dropout(config.classifier_dropout_prob , inplace=__UpperCAmelCase) a : Optional[int] = nn.Linear(__UpperCAmelCase , config.num_labels) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__UpperCAmelCase) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=__UpperCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def __snake_case ( self : Dict , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , ): a : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict a : Dict = self.mobilenet_va(__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , return_dict=__UpperCAmelCase) a : List[Any] = outputs.pooler_output if return_dict else outputs[1] a : List[Any] = self.classifier(self.dropout(__UpperCAmelCase)) a : Any = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: a : List[str] = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): a : Any = "single_label_classification" else: a : Optional[Any] = "multi_label_classification" if self.config.problem_type == "regression": a : Union[str, Any] = MSELoss() if self.num_labels == 1: a : Tuple = loss_fct(logits.squeeze() , labels.squeeze()) else: a : Dict = loss_fct(__UpperCAmelCase , __UpperCAmelCase) elif self.config.problem_type == "single_label_classification": a : Optional[Any] = CrossEntropyLoss() a : Tuple = loss_fct(logits.view(-1 , self.num_labels) , labels.view(-1)) elif self.config.problem_type == "multi_label_classification": a : Dict = BCEWithLogitsLoss() a : List[Any] = loss_fct(__UpperCAmelCase , __UpperCAmelCase) if not return_dict: a : int = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=__UpperCAmelCase , logits=__UpperCAmelCase , hidden_states=outputs.hidden_states , )
226
0
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } _UpperCAmelCase = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } _UpperCAmelCase = { 'facebook/blenderbot_small-90M': 512, } class UpperCAmelCase ( lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ = BlenderbotSmallTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase="<|endoftext|>" , lowercase="<|endoftext|>" , lowercase="<|endoftext|>" , lowercase=False , lowercase=True , **lowercase , ): """simple docstring""" super().__init__( ByteLevelBPETokenizer( vocab=UpperCamelCase__ , merges=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ , ) , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , **UpperCamelCase__ , ) A_ : List[str] = add_prefix_space def lowerCAmelCase_ ( self , lowercase , lowercase=None ): """simple docstring""" A_ : Any = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowerCAmelCase_ ( self , lowercase , lowercase = None ): """simple docstring""" A_ : Tuple = [self.sep_token_id] A_ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
140
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Dict: # Initialise PyTorch model lowerCamelCase : Any = TaConfig.from_json_file(_SCREAMING_SNAKE_CASE ) print(f'''Building PyTorch model from configuration: {config}''' ) lowerCamelCase : str = TaForConditionalGeneration(_SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_ta(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) SCREAMING_SNAKE_CASE__ : str = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
48
0
"""simple docstring""" from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record a_ = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ a_ = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ a_ = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def __lowercase ( snake_case_ : int ,snake_case_ : Optional[int] ) ->Union[str, Any]: '''simple docstring''' return float((preds == labels).mean() ) def __lowercase ( snake_case_ : int ,snake_case_ : Any ,snake_case_ : Union[str, Any]="binary" ) ->Optional[int]: '''simple docstring''' __A : List[Any] = simple_accuracy(snake_case_ ,snake_case_ ) __A : List[str] = float(fa_score(y_true=snake_case_ ,y_pred=snake_case_ ,average=snake_case_ ) ) return { "accuracy": acc, "f1": fa, } def __lowercase ( snake_case_ : List[Any] ,snake_case_ : List[str] ) ->Tuple: '''simple docstring''' __A : int = {} for id_pred, label in zip(snake_case_ ,snake_case_ ): __A : int = F"""{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}""" __A : List[str] = id_pred['''prediction'''] if question_id in question_map: question_map[question_id].append((pred, label) ) else: __A : Tuple = [(pred, label)] __A : int = [], [] for question, preds_labels in question_map.items(): __A : List[Any] = zip(*snake_case_ ) __A : str = fa_score(y_true=snake_case_ ,y_pred=snake_case_ ,average='''macro''' ) fas.append(snake_case_ ) __A : Union[str, Any] = int(sum(pred == label for pred, label in preds_labels ) == len(snake_case_ ) ) ems.append(snake_case_ ) __A : Union[str, Any] = float(sum(snake_case_ ) / len(snake_case_ ) ) __A : Tuple = sum(snake_case_ ) / len(snake_case_ ) __A : Dict = float(fa_score(y_true=snake_case_ ,y_pred=[id_pred['''prediction'''] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): """simple docstring""" def UpperCamelCase__( self ): '''simple docstring''' if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if not self.config_name == '''record''' and not self.config_name == '''multirc''' else None , ) def UpperCamelCase__( self ): '''simple docstring''' if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "prediction_text": datasets.Value('''string''' ), }, "references": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "answers": datasets.Sequence(datasets.Value('''string''' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('''int64''' ), "paragraph": datasets.Value('''int64''' ), "question": datasets.Value('''int64''' ), }, "prediction": datasets.Value('''int64''' ), }, "references": datasets.Value('''int64''' ), } else: return { "predictions": datasets.Value('''int64''' ), "references": datasets.Value('''int64''' ), } def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(__lowerCamelCase , __lowerCamelCase )} elif self.config_name == "cb": return acc_and_fa(__lowerCamelCase , __lowerCamelCase , fa_avg='''macro''' ) elif self.config_name == "record": __A : Any = [ { '''qas''': [ {'''id''': ref['''idx''']['''query'''], '''answers''': [{'''text''': ans} for ans in ref['''answers''']]} for ref in references ] } ] __A : str = {pred['''idx''']['''query''']: pred['''prediction_text'''] for pred in predictions} return evaluate_record(__lowerCamelCase , __lowerCamelCase )[0] elif self.config_name == "multirc": return evaluate_multirc(__lowerCamelCase , __lowerCamelCase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(__lowerCamelCase , __lowerCamelCase )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' )
365
"""simple docstring""" from math import factorial def __lowercase ( snake_case_ : int ,snake_case_ : int ) ->int: '''simple docstring''' if n < k or k < 0: raise ValueError('''Please enter positive integers for n and k where n >= k''' ) return factorial(snake_case_ ) // (factorial(snake_case_ ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", f'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", f'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", f'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
291
0
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCAmelCase = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCAmelCase = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCAmelCase = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCAmelCase = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _a ( datasets.Metric ): def lowerCamelCase_ ( self: Optional[Any] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase_: Optional[Any] ) -> List[str]: """simple docstring""" import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def lowerCamelCase_ ( self: int , UpperCamelCase_: Any , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Union[str, Any]=0.9 , UpperCamelCase_: List[Any]=3 , UpperCamelCase_: Tuple=0.5 ) -> List[Any]: """simple docstring""" if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase__ = [ meteor_score.single_meteor_score( word_tokenize(lowerCAmelCase_ ) , word_tokenize(lowerCAmelCase_ ) , alpha=lowerCAmelCase_ , beta=lowerCAmelCase_ , gamma=lowerCAmelCase_ ) for ref, pred in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ] else: lowercase__ = [ meteor_score.single_meteor_score(lowerCAmelCase_ , lowerCAmelCase_ , alpha=lowerCAmelCase_ , beta=lowerCAmelCase_ , gamma=lowerCAmelCase_ ) for ref, pred in zip(lowerCAmelCase_ , lowerCAmelCase_ ) ] return {"meteor": np.mean(lowerCAmelCase_ )}
110
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = StableDiffusionInstructPixaPixPipeline a_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS a_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Optional[int] ) -> Optional[int]: torch.manual_seed(0 ) __lowerCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) __lowerCAmelCase = PNDMScheduler(skip_prk_steps=lowerCAmelCase_ ) torch.manual_seed(0 ) __lowerCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) __lowerCAmelCase = CLIPTextModel(lowerCAmelCase_ ) __lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ) -> Dict: __lowerCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCAmelCase = Image.fromarray(np.uinta(lowerCAmelCase_ ) ).convert('RGB' ) if str(lowerCAmelCase_ ).startswith('mps' ): __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) else: __lowerCAmelCase = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Tuple ) -> List[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Dict: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = 'french fries' __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ , negative_prompt=lowerCAmelCase_ ) __lowerCAmelCase = output.images __lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = [inputs['prompt']] * 2 __lowerCAmelCase = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ).unsqueeze(0 ).to(lowerCAmelCase_ ) __lowerCAmelCase = image / 2 + 0.5 __lowerCAmelCase = image.permute(0 , 3 , 1 , 2 ) __lowerCAmelCase = image.repeat(2 , 1 , 1 , 1 ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Dict ) -> Optional[Any]: __lowerCAmelCase = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe.to(lowerCAmelCase_ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = self.get_dummy_inputs(lowerCAmelCase_ ) __lowerCAmelCase = sd_pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1] __lowerCAmelCase = [round(lowerCAmelCase_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(lowerCAmelCase_ ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = self.get_dummy_components() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase_ ) __lowerCAmelCase = VaeImageProcessor(do_resize=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) __lowerCAmelCase = pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) )[0] __lowerCAmelCase = components['vae'] __lowerCAmelCase = self.get_dummy_inputs_by_type(lowerCAmelCase_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCAmelCase = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCAmelCase = pipe(**lowerCAmelCase_ )[0] __lowerCAmelCase = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase_ , 1e-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : int ) -> Optional[int]: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any]=0 ) -> Any: __lowerCAmelCase = torch.manual_seed(lowerCAmelCase_ ) __lowerCAmelCase = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCAmelCase = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase ( self : List[Any] ) -> str: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Tuple ) -> List[str]: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[Any] ) -> Dict: __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ ) __lowerCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ).images __lowerCAmelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase ( self : Optional[int] ) -> int: __lowerCAmelCase = 0 def callback_fn(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : torch.FloatTensor ) -> None: __lowerCAmelCase = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __lowerCAmelCase = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) __lowerCAmelCase = latents[0, -3:, -3:, -1] __lowerCAmelCase = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __lowerCAmelCase = False __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = self.get_inputs() pipe(**lowerCAmelCase_ , callback=lowerCAmelCase_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase ( self : Optional[int] ) -> Any: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=lowerCAmelCase_ , torch_dtype=torch.floataa ) __lowerCAmelCase = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCAmelCase = self.get_inputs() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCAmelCase = inputs['image'].resize((5_0_4, 5_0_4) ) __lowerCAmelCase = 'timbrooks/instruct-pix2pix' __lowerCAmelCase = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase_ , safety_checker=lowerCAmelCase_ , ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) pipe.enable_attention_slicing() __lowerCAmelCase = pipe(**lowerCAmelCase_ ) __lowerCAmelCase = output.images[0] __lowerCAmelCase = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) __lowerCAmelCase = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
284
0
'''simple docstring''' import os import time import numpy as np import onnxruntime as ort _A : Optional[Any] ='''1''' _A : Any ='''0''' _A : Union[str, Any] ='''1''' _A : Tuple =ort.SessionOptions() _A : List[Any] =ort.GraphOptimizationLevel.ORT_DISABLE_ALL print('''Create inference session...''') _A : int =['''TensorrtExecutionProvider''', '''CUDAExecutionProvider'''] _A : Tuple =ort.InferenceSession('''model.onnx''', sess_options=sess_opt, providers=execution_provider) _A : List[Any] =ort.RunOptions() _A : List[Any] =128 _A : Dict =1 _A : List[str] =np.ones((batch, sequence), dtype=np.intaa) _A : Union[str, Any] =np.ones((batch, sequence), dtype=np.intaa) _A : str =np.ones((batch, sequence), dtype=np.intaa) print('''Warm up phase...''') sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('''Start inference...''') _A : int =time.time() _A : List[str] =2_000 _A : List[str] ={} for iter in range(max_iters): _A : List[Any] =sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('''Average Inference Time = {:.3f} ms'''.format((time.time() - start_time) * 1_000 / max_iters))
353
'''simple docstring''' import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient _A : str =WebClient(token=os.environ['''CI_SLACK_BOT_TOKEN''']) def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Tuple: lowerCamelCase__ : int = test_results.split(""" """ ) lowerCamelCase__ : Optional[int] = 0 lowerCamelCase__ : Any = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. lowerCamelCase__ : Union[str, Any] = expressions[-2] if """=""" in expressions[-1] else expressions[-1] for i, expression in enumerate(UpperCamelCase ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Tuple: lowerCamelCase__ : Optional[Any] = {} lowerCamelCase__ : int = None lowerCamelCase__ : Optional[int] = False for line in failures_short_lines.split("""\n""" ): if re.search(r"""_ \[doctest\]""" , UpperCamelCase ): lowerCamelCase__ : Dict = True lowerCamelCase__ : Optional[int] = line.split(""" """ )[2] elif in_error and not line.split(""" """ )[0].isdigit(): lowerCamelCase__ : List[str] = line lowerCamelCase__ : int = False return failures class _lowercase : def __init__( self: Tuple , UpperCamelCase__: str , UpperCamelCase__: Dict ): lowerCamelCase__ : Union[str, Any] = title lowerCamelCase__ : Tuple = doc_test_results["""time_spent"""].split(""",""" )[0] lowerCamelCase__ : Union[str, Any] = doc_test_results["""success"""] lowerCamelCase__ : List[Any] = doc_test_results["""failures"""] lowerCamelCase__ : List[str] = self.n_success + self.n_failures # Failures and success of the modeling tests lowerCamelCase__ : str = doc_test_results @property def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : Union[str, Any] = [self._time_spent] lowerCamelCase__ : Tuple = 0 for time in time_spent: lowerCamelCase__ : Tuple = time.split(""":""" ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(UpperCamelCase__ ) == 1: lowerCamelCase__ : Tuple = [0, 0, time_parts[0]] lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : List[Any] = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 3_600 + minutes * 60 + seconds lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : List[Any] = total_secs // 3_600, (total_secs % 3_600) // 60, total_secs % 60 return F'''{int(UpperCamelCase__ )}h{int(UpperCamelCase__ )}m{int(UpperCamelCase__ )}s''' @property def lowerCamelCase_ ( self: Dict ): return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def lowerCamelCase_ ( self: Any ): return { "type": "section", "text": { "type": "plain_text", "text": F'''🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.''', "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } @property def lowerCamelCase_ ( self: Any ): return { "type": "section", "text": { "type": "plain_text", "text": ( F'''There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in''' F''' {self.time}.''' ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } @property def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Any = 40 lowerCamelCase__ : List[str] = {k: v["""failed"""] for k, v in doc_test_results.items() if isinstance(UpperCamelCase__ , UpperCamelCase__ )} lowerCamelCase__ : List[Any] = """""" for category, failures in category_failures.items(): if len(UpperCamelCase__ ) == 0: continue if report != "": report += "\n\n" report += F'''*{category} failures*:'''.ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(UpperCamelCase__ ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": F'''The following examples had failures:\n\n\n{report}\n''', }, } @property def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Union[str, Any] = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(UpperCamelCase__ ) @staticmethod def lowerCamelCase_ ( ): lowerCamelCase__ : List[Any] = [ { """type""": """section""", """text""": { """type""": """plain_text""", """text""": """There was an issue running the tests.""", }, """accessory""": { """type""": """button""", """text""": {"""type""": """plain_text""", """text""": """Check Action results""", """emoji""": True}, """url""": F'''https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } ] print("""Sending the following payload""" ) print(json.dumps({"""blocks""": json.loads(UpperCamelCase__ )} ) ) client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , text="""There was an issue running the tests.""" , blocks=UpperCamelCase__ , ) def lowerCamelCase_ ( self: Any ): print("""Sending the following payload""" ) print(json.dumps({"""blocks""": json.loads(self.payload )} ) ) lowerCamelCase__ : Any = F'''{self.n_failures} failures out of {self.n_tests} tests,''' if self.n_failures else """All tests passed.""" lowerCamelCase__ : List[str] = client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , blocks=self.payload , text=UpperCamelCase__ , ) def lowerCamelCase_ ( self: Any , UpperCamelCase__: str , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : Optional[Any] = """""" for key, value in failures.items(): lowerCamelCase__ : int = value[:200] + """ [Truncated]""" if len(UpperCamelCase__ ) > 250 else value failures_text += F'''*{key}*\n_{value}_\n\n''' lowerCamelCase__ : Tuple = job_name lowerCamelCase__ : Union[str, Any] = {"""type""": """section""", """text""": {"""type""": """mrkdwn""", """text""": text}} if job_link is not None: lowerCamelCase__ : Union[str, Any] = { """type""": """button""", """text""": {"""type""": """plain_text""", """text""": """GitHub Action job""", """emoji""": True}, """url""": job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def lowerCamelCase_ ( self: Tuple ): if self.thread_ts is None: raise ValueError("""Can only post reply if a post has been made.""" ) lowerCamelCase__ : int = self.doc_test_results.pop("""job_link""" ) self.doc_test_results.pop("""failures""" ) self.doc_test_results.pop("""success""" ) self.doc_test_results.pop("""time_spent""" ) lowerCamelCase__ : List[Any] = sorted(self.doc_test_results.items() , key=lambda UpperCamelCase__ : t[0] ) for job, job_result in sorted_dict: if len(job_result["""failures"""] ): lowerCamelCase__ : Union[str, Any] = F'''*Num failures* :{len(job_result['failed'] )} \n''' lowerCamelCase__ : Union[str, Any] = job_result["""failures"""] lowerCamelCase__ : int = self.get_reply_blocks(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , text=UpperCamelCase__ ) print("""Sending the following reply""" ) print(json.dumps({"""blocks""": blocks} ) ) client.chat_postMessage( channel=os.environ["""CI_SLACK_CHANNEL_ID_DAILY"""] , text=F'''Results for {job}''' , blocks=UpperCamelCase__ , thread_ts=self.thread_ts["""ts"""] , ) time.sleep(1 ) def SCREAMING_SNAKE_CASE_ () -> Tuple: lowerCamelCase__ : Any = os.environ["""GITHUB_RUN_ID"""] lowerCamelCase__ : List[Any] = f'''https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100''' lowerCamelCase__ : Optional[int] = requests.get(UpperCamelCase ).json() lowerCamelCase__ : List[Any] = {} try: jobs.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) lowerCamelCase__ : Any = math.ceil((result["""total_count"""] - 100) / 100 ) for i in range(UpperCamelCase ): lowerCamelCase__ : List[Any] = requests.get(url + f'''&page={i + 2}''' ).json() jobs.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) return jobs except Exception as e: print("""Unknown error, could not fetch links.""" , UpperCamelCase ) return {} def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Any: lowerCamelCase__ : int = {} if os.path.exists(UpperCamelCase ): lowerCamelCase__ : List[str] = os.listdir(UpperCamelCase ) for file in files: try: with open(os.path.join(UpperCamelCase , UpperCamelCase ) , encoding="""utf-8""" ) as f: lowerCamelCase__ : List[Any] = f.read() except UnicodeDecodeError as e: raise ValueError(f'''Could not open {os.path.join(UpperCamelCase , UpperCamelCase )}.''' ) from e return _artifact def SCREAMING_SNAKE_CASE_ () -> Optional[Any]: class _lowercase : def __init__( self: Tuple , UpperCamelCase__: str ): lowerCamelCase__ : Any = name lowerCamelCase__ : Union[str, Any] = [] def __str__( self: int ): return self.name def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: str ): self.paths.append({"""name""": self.name, """path""": path} ) lowerCamelCase__ : Dict[str, Artifact] = {} lowerCamelCase__ : List[str] = filter(os.path.isdir , os.listdir() ) for directory in directories: lowerCamelCase__ : Union[str, Any] = directory if artifact_name not in _available_artifacts: lowerCamelCase__ : Optional[int] = Artifact(UpperCamelCase ) _available_artifacts[artifact_name].add_path(UpperCamelCase ) return _available_artifacts if __name__ == "__main__": _A : Any =get_job_links() _A : str =retrieve_available_artifacts() _A : int =collections.OrderedDict( [ ('''*.py''', '''API Examples'''), ('''*.md''', '''MD Examples'''), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' _A : Union[str, Any] ={ v: { '''failed''': [], '''failures''': {}, } for v in docs.values() } # Link to the GitHub Action job _A : Union[str, Any] =github_actions_job_links.get('''run_doctests''') _A : Any =available_artifacts['''doc_tests_gpu_test_reports'''].paths[0] _A : Dict =retrieve_artifact(artifact_path['''name''']) if "stats" in artifact: _A , _A , _A : Optional[int] =handle_test_results(artifact['''stats''']) _A : Union[str, Any] =failed _A : int =success _A : Optional[int] =time_spent[1:-1] + ''', ''' _A : Any =extract_first_line_failure(artifact['''failures_short''']) for line in artifact["summary_short"].split('''\n'''): if re.search('''FAILED''', line): _A : List[Any] =line.replace('''FAILED ''', '''''') _A : Any =line.split()[0].replace('''\n''', '''''') if "::" in line: _A , _A : Any =line.split('''::''') else: _A , _A : Tuple =line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): _A : str =docs[file_regex] doc_test_results[category]["failed"].append(test) _A : str =all_failures[test] if test in all_failures else '''N/A''' _A : Tuple =failure break _A : Union[str, Any] =Message('''🤗 Results of the doc tests.''', doc_test_results) message.post() message.post_reply()
129
0
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowerCAmelCase_ ( __a ) -> str: """simple docstring""" lowerCamelCase__: List[Any] =[ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "_float_tensor", "decoder.output_projection.weight", ] for k in ignore_keys: state_dict.pop(lowerCamelCase__ , lowerCamelCase__ ) def lowerCAmelCase_ ( __a ) -> List[str]: """simple docstring""" lowerCamelCase__: Union[str, Any] =emb.weight.shape lowerCamelCase__: Union[str, Any] =nn.Linear(lowerCamelCase__ , lowerCamelCase__ , bias=lowerCamelCase__ ) lowerCamelCase__: Dict =emb.weight.data return lin_layer def lowerCAmelCase_ ( __a , __a="facebook/mbart-large-en-ro" , __a=False , __a=False ) -> Tuple: """simple docstring""" lowerCamelCase__: List[Any] =torch.load(lowerCamelCase__ , map_location="cpu" )["model"] remove_ignore_keys_(lowerCamelCase__ ) lowerCamelCase__: List[Any] =state_dict["encoder.embed_tokens.weight"].shape[0] lowerCamelCase__: Optional[Any] =MBartConfig.from_pretrained(lowerCamelCase__ , vocab_size=lowerCamelCase__ ) if mbart_aa and finetuned: lowerCamelCase__: List[str] ="relu" lowerCamelCase__: Union[str, Any] =state_dict["decoder.embed_tokens.weight"] lowerCamelCase__: Tuple =MBartForConditionalGeneration(lowerCamelCase__ ) model.model.load_state_dict(lowerCamelCase__ ) if finetuned: lowerCamelCase__: Optional[Any] =make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( "fairseq_path", type=str, help="bart.large, bart.large.cnn or a path to a model.pt on local filesystem." ) parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument( "--hf_config", default="facebook/mbart-large-cc25", type=str, help="Which huggingface architecture to use: mbart-large", ) parser.add_argument("--mbart_50", action="store_true", help="whether the model is mMART-50 checkpoint") parser.add_argument("--finetuned", action="store_true", help="whether the model is a fine-tuned checkpoint") __A = parser.parse_args() __A = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
10
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCAmelCase__ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''BartphoTokenizer'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
130
0
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { '''linear''': get_linear_schedule_with_warmup, '''cosine''': get_cosine_schedule_with_warmup, '''cosine_w_restarts''': get_cosine_with_hard_restarts_schedule_with_warmup, '''polynomial''': get_polynomial_decay_schedule_with_warmup, '''constant''': get_constant_schedule, '''constant_w_warmup''': get_constant_schedule_with_warmup, } class lowerCamelCase_ ( __snake_case ): def __init__( self : int , _A : Union[str, Any]=None , _A : Dict=None , *_A : Any , **_A : str ): '''simple docstring''' super().__init__(*UpperCamelCase__ , **UpperCamelCase__ ) if config is None: assert isinstance(self.model , UpperCamelCase__ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" f""" {self.model.__class__}""" ) UpperCAmelCase__ : Dict = self.model.config else: UpperCAmelCase__ : str = config UpperCAmelCase__ : Any = data_args UpperCAmelCase__ : Tuple = self.config.tgt_vocab_size if isinstance(self.config , UpperCamelCase__ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( f"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: UpperCAmelCase__ : Dict = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss UpperCAmelCase__ : List[str] = label_smoothed_nll_loss def lowercase_ ( self : Dict , _A : int ): '''simple docstring''' if self.optimizer is None: UpperCAmelCase__ : Optional[Any] = ["bias", "LayerNorm.weight"] UpperCAmelCase__ : Tuple = [ { "params": [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], "weight_decay": self.args.weight_decay, }, { "params": [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], "weight_decay": 0.0, }, ] UpperCAmelCase__ : List[str] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: UpperCAmelCase__ : Dict = Adafactor UpperCAmelCase__ : Optional[Any] = {"scale_parameter": False, "relative_step": False} else: UpperCAmelCase__ : List[str] = AdamW UpperCAmelCase__ : Any = { "betas": (self.args.adam_betaa, self.args.adam_betaa), "eps": self.args.adam_epsilon, } UpperCAmelCase__ : Tuple = self.args.learning_rate if self.sharded_ddp: UpperCAmelCase__ : List[str] = OSS( params=UpperCamelCase__ , optim=UpperCamelCase__ , **UpperCamelCase__ , ) else: UpperCAmelCase__ : Union[str, Any] = optimizer_cls(UpperCamelCase__ , **UpperCamelCase__ ) if self.lr_scheduler is None: UpperCAmelCase__ : Optional[Any] = self._get_lr_scheduler(UpperCamelCase__ ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def lowercase_ ( self : Optional[Any] , _A : Any ): '''simple docstring''' UpperCAmelCase__ : int = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": UpperCAmelCase__ : str = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": UpperCAmelCase__ : Any = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: UpperCAmelCase__ : int = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=UpperCamelCase__ ) return scheduler def lowercase_ ( self : List[str] ): '''simple docstring''' if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def lowercase_ ( self : str , _A : Optional[Any] , _A : str , _A : List[str] ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token UpperCAmelCase__ : int = model(**UpperCamelCase__ , use_cache=UpperCamelCase__ )[0] UpperCAmelCase__ : int = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models UpperCAmelCase__ : Any = model(**UpperCamelCase__ , labels=UpperCamelCase__ , use_cache=UpperCamelCase__ )[:2] else: # compute label smoothed loss UpperCAmelCase__ : Any = model(**UpperCamelCase__ , use_cache=UpperCamelCase__ )[0] UpperCAmelCase__ : str = torch.nn.functional.log_softmax(UpperCamelCase__ , dim=-1 ) UpperCAmelCase__ : int = self.loss_fn(UpperCamelCase__ , UpperCamelCase__ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def lowercase_ ( self : Optional[Any] , _A : Any , _A : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = inputs.pop('''labels''' ) UpperCAmelCase__ : str = self._compute_loss(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return loss def lowercase_ ( self : List[str] , _A : nn.Module , _A : Dict[str, Union[torch.Tensor, Any]] , _A : bool , _A : Optional[List[str]] = None , ): '''simple docstring''' UpperCAmelCase__ : Dict = self._prepare_inputs(UpperCamelCase__ ) UpperCAmelCase__ : Optional[int] = { "max_length": self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, "num_beams": self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: UpperCAmelCase__ : int = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **UpperCamelCase__ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: UpperCAmelCase__ : Tuple = self._pad_tensors_to_max_len(UpperCamelCase__ , gen_kwargs['''max_length'''] ) UpperCAmelCase__ : Optional[Any] = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data UpperCAmelCase__ : int = self._compute_loss(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) UpperCAmelCase__ : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) UpperCAmelCase__ : Any = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: UpperCAmelCase__ : str = self._pad_tensors_to_max_len(UpperCamelCase__ , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def lowercase_ ( self : List[Any] , _A : Dict , _A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' f""" padded to `max_length`={max_length}""" ) UpperCAmelCase__ : str = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) UpperCAmelCase__ : List[Any] = tensor return padded_tensor
370
'''simple docstring''' from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def a__ ( lowerCAmelCase__ ) -> None: UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = analyze_text(lowerCAmelCase__ ) UpperCAmelCase__ : List[Any] = list(''' ''' + ascii_lowercase ) # what is our total sum of probabilities. UpperCAmelCase__ : str = sum(single_char_strings.values() ) # one length string UpperCAmelCase__ : int = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: UpperCAmelCase__ : Optional[int] = single_char_strings[ch] UpperCAmelCase__ : int = my_str / all_sum my_fir_sum += prob * math.loga(lowerCAmelCase__ ) # entropy formula. # print entropy print(F"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string UpperCAmelCase__ : str = sum(two_char_strings.values() ) UpperCAmelCase__ : Optional[Any] = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: UpperCAmelCase__ : Optional[int] = cha + cha if sequence in two_char_strings: UpperCAmelCase__ : Dict = two_char_strings[sequence] UpperCAmelCase__ : Optional[int] = int(lowerCAmelCase__ ) / all_sum my_sec_sum += prob * math.loga(lowerCAmelCase__ ) # print second entropy print(F"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(F"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def a__ ( lowerCAmelCase__ ) -> tuple[dict, dict]: UpperCAmelCase__ : Union[str, Any] = Counter() # type: ignore UpperCAmelCase__ : Tuple = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(lowerCAmelCase__ ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def a__ ( ) -> Tuple: import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
299
0
"""simple docstring""" def A_ ( _lowercase ): '''simple docstring''' snake_case_ :int = (1 + 24 * n) ** 0.5 return ((1 + root) / 6) % 1 == 0 def A_ ( _lowercase = 5000 ): '''simple docstring''' snake_case_ :Any = [(i * (3 * i - 1)) // 2 for i in range(1, _lowercase )] for i, pentagonal_i in enumerate(_lowercase ): for j in range(_lowercase, len(_lowercase ) ): snake_case_ :List[str] = pentagonal_nums[j] snake_case_ :Dict = pentagonal_i + pentagonal_j snake_case_ :str = pentagonal_j - pentagonal_i if is_pentagonal(_lowercase ) and is_pentagonal(_lowercase ): return b return -1 if __name__ == "__main__": print(F"""{solution() = }""")
66
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A: Optional[int] = logging.get_logger(__name__) A: Optional[int] = torch.device("cpu") def _snake_case ( ): UpperCAmelCase : List[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCAmelCase : Tuple = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return im def _snake_case ( UpperCamelCase : int ): if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0, 8.8_6_8_5e-0_1, 2.4_3_6_0e-0_1] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9_6_3_6e-0_1, 2.3_4_7_8e-0_1, -1.6_9_6_3e0_0, -1.7_3_8_1e0_0, -8.6_3_3_7e-0_1] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2_7_6_8e-0_1, -4.7_4_2_9e-0_1, -1.0_8_9_7e0_0, -1.0_2_4_8e0_0, 3.5_5_2_3e-0_2] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5_3_3_0e-0_1, 2.4_2_1_1e-0_1, -6.0_1_8_5e-0_1, -8.2_7_8_9e-0_1, -6.0_4_4_6e-0_2] ) def _snake_case ( UpperCamelCase : Optional[Any] , UpperCamelCase : List[str] , UpperCamelCase : str ): UpperCAmelCase : int = dct.pop(UpperCamelCase ) UpperCAmelCase : Any = val def _snake_case ( UpperCamelCase : Union[str, Any] ): UpperCAmelCase : Optional[int] = [] for k in state_dict.keys(): UpperCAmelCase : Optional[Any] = k if ".pwconv" in k: UpperCAmelCase : int = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: UpperCAmelCase : Tuple = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: UpperCAmelCase : List[Any] = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: UpperCAmelCase : Any = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: UpperCAmelCase : int = k_new.split(""".""" ) if ls[2].isdigit(): UpperCAmelCase : List[Any] = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: UpperCAmelCase : Any = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def _snake_case ( UpperCamelCase : int , UpperCamelCase : int , UpperCamelCase : Optional[int] ): UpperCAmelCase : List[Any] = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase : Optional[Any] = 1000 UpperCAmelCase : Tuple = """huggingface/label-files""" UpperCAmelCase : List[str] = """imagenet-1k-id2label.json""" UpperCAmelCase : Dict = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase : Dict = {int(UpperCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase : str = idalabel UpperCAmelCase : Dict = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": UpperCAmelCase : Any = [3, 3, 6, 4] UpperCAmelCase : List[str] = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": UpperCAmelCase : Dict = [3, 3, 9, 6] UpperCAmelCase : Union[str, Any] = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": UpperCAmelCase : int = [4, 3, 10, 5] UpperCAmelCase : Optional[int] = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": UpperCAmelCase : Union[str, Any] = [4, 4, 12, 6] UpperCAmelCase : List[Any] = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): UpperCAmelCase : List[str] = torch.hub.load_state_dict_from_url(UpperCamelCase , map_location="""cpu""" , check_hash=UpperCamelCase ) else: UpperCAmelCase : Any = torch.load(UpperCamelCase , map_location="""cpu""" ) UpperCAmelCase : Optional[Any] = checkpoint UpperCAmelCase : Dict = create_rename_keys(UpperCamelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # load HuggingFace model UpperCAmelCase : List[Any] = SwiftFormerForImageClassification(UpperCamelCase ).eval() hf_model.load_state_dict(UpperCamelCase ) # prepare test inputs UpperCAmelCase : Dict = prepare_img() UpperCAmelCase : Tuple = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) UpperCAmelCase : Optional[int] = processor(images=UpperCamelCase , return_tensors="""pt""" ) # compare outputs from both models UpperCAmelCase : Optional[int] = get_expected_output(UpperCamelCase ) UpperCAmelCase : List[str] = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1000] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase , atol=1e-3 ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) print(F"Saving model {swiftformer_name} to {pytorch_dump_folder_path}" ) hf_model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": A: List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--swiftformer_name", default="swiftformer_xs", choices=["swiftformer_xs", "swiftformer_s", "swiftformer_l1", "swiftformer_l3"], type=str, help="Name of the SwiftFormer model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default="./converted_outputs/", type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument("--original_ckpt", default=None, type=str, help="Path to the original model checkpoint.") A: str = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
109
0
def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :Tuple ) -> int: if not head: return True # split the list to two parts __lowerCAmelCase , __lowerCAmelCase : Optional[int] = head.next, head while fast and fast.next: __lowerCAmelCase : Any = fast.next.next __lowerCAmelCase : Tuple = slow.next __lowerCAmelCase : Union[str, Any] = slow.next __lowerCAmelCase : Optional[int] = None # Don't forget here! But forget still works! # reverse the second part __lowerCAmelCase : Tuple = None while second: __lowerCAmelCase : int = second.next __lowerCAmelCase : Optional[Any] = node __lowerCAmelCase : List[str] = second __lowerCAmelCase : Optional[Any] = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False __lowerCAmelCase : Any = node.next __lowerCAmelCase : Optional[int] = head.next return True def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :List[Any] ) -> List[Any]: if not head or not head.next: return True # 1. Get the midpoint (slow) __lowerCAmelCase : str = head while fast and fast.next: __lowerCAmelCase , __lowerCAmelCase : List[Any] = fast.next.next, slow.next # 2. Push the second half into the stack __lowerCAmelCase : Optional[Any] = [slow.val] while slow.next: __lowerCAmelCase : List[str] = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False __lowerCAmelCase : Optional[int] = cur.next return True def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :Optional[int] ) -> Dict: if not head or not head.next: return True __lowerCAmelCase : List[str] = {} __lowerCAmelCase : str = 0 while head: if head.val in d: d[head.val].append(SCREAMING_SNAKE_CASE ) else: __lowerCAmelCase : Optional[Any] = [pos] __lowerCAmelCase : Any = head.next pos += 1 __lowerCAmelCase : List[str] = pos - 1 __lowerCAmelCase : int = 0 for v in d.values(): if len(SCREAMING_SNAKE_CASE ) % 2 != 0: middle += 1 else: __lowerCAmelCase : int = 0 for i in range(0 , len(SCREAMING_SNAKE_CASE ) ): if v[i] + v[len(SCREAMING_SNAKE_CASE ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
232
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase = '▁' _UpperCAmelCase = {'vocab_file': 'spiece.model'} _UpperCAmelCase = { 'vocab_file': {'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'} } _UpperCAmelCase = { 'google/pegasus-xsum': 512, } _UpperCAmelCase = logging.get_logger(__name__) class snake_case_ ( __lowercase ): A_ = VOCAB_FILES_NAMES A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = ['input_ids', 'attention_mask'] def __init__( self : List[Any] , _snake_case : Union[str, Any] , _snake_case : Union[str, Any]="<pad>" , _snake_case : int="</s>" , _snake_case : Any="<unk>" , _snake_case : Union[str, Any]="<mask_2>" , _snake_case : Any="<mask_1>" , _snake_case : Optional[int]=None , _snake_case : List[str]=103 , _snake_case : Optional[Dict[str, Any]] = None , **_snake_case : Optional[int] , )->None: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = offset if additional_special_tokens is not None: if not isinstance(_snake_case , _snake_case ): raise TypeError( F'''additional_special_tokens should be of type {type(_snake_case )}, but is''' F''' {type(_snake_case )}''' ) __lowerCAmelCase : List[str] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F'''<unk_{i}>''' for i in range(len(_snake_case ) , self.offset - 1 ) ] if len(set(_snake_case ) ) != len(_snake_case ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" F''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) __lowerCAmelCase : Dict = additional_special_tokens_extended else: __lowerCAmelCase : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F'''<unk_{i}>''' for i in range(2 , self.offset )] __lowerCAmelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=_snake_case , unk_token=_snake_case , mask_token=_snake_case , pad_token=_snake_case , mask_token_sent=_snake_case , offset=_snake_case , additional_special_tokens=_snake_case , sp_model_kwargs=self.sp_model_kwargs , **_snake_case , ) __lowerCAmelCase : Optional[Any] = mask_token_sent __lowerCAmelCase : Any = vocab_file __lowerCAmelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_snake_case ) # add special tokens to encoder dict __lowerCAmelCase : Dict[int, str] = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) __lowerCAmelCase : Dict[str, int] = {v: k for k, v in self.encoder.items()} @property def UpperCAmelCase__ ( self : str )->int: '''simple docstring''' return len(self.sp_model ) + self.offset def UpperCAmelCase__ ( self : Dict )->Dict[str, int]: '''simple docstring''' __lowerCAmelCase : Tuple = {self.convert_ids_to_tokens(_snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] )->str: '''simple docstring''' __lowerCAmelCase : Optional[Any] = self.__dict__.copy() __lowerCAmelCase : Union[str, Any] = None return state def __setstate__( self : Any , _snake_case : str )->Any: '''simple docstring''' __lowerCAmelCase : str = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __lowerCAmelCase : Any = {} __lowerCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCAmelCase__ ( self : Dict , _snake_case : str )->List[str]: '''simple docstring''' return self.sp_model.encode(_snake_case , out_type=_snake_case ) def UpperCAmelCase__ ( self : Tuple , _snake_case : str )->int: '''simple docstring''' if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] __lowerCAmelCase : Any = self.sp_model.piece_to_id(_snake_case ) return sp_id + self.offset def UpperCAmelCase__ ( self : List[Any] , _snake_case : int )->str: '''simple docstring''' if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: __lowerCAmelCase : Optional[int] = self.sp_model.IdToPiece(index - self.offset ) return token def UpperCAmelCase__ ( self : Union[str, Any] , _snake_case : Optional[int] )->List[str]: '''simple docstring''' __lowerCAmelCase : Any = [] __lowerCAmelCase : Dict = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(_snake_case ) + token __lowerCAmelCase : int = [] else: current_sub_tokens.append(_snake_case ) out_string += self.sp_model.decode(_snake_case ) return out_string.strip() def UpperCAmelCase__ ( self : Optional[int] , _snake_case : Dict=False )->int: '''simple docstring''' return 1 def UpperCAmelCase__ ( self : Tuple , _snake_case : Tuple )->str: '''simple docstring''' __lowerCAmelCase : List[Any] = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def UpperCAmelCase__ ( self : List[str] , _snake_case : List , _snake_case : Optional[List] = None , _snake_case : bool = False )->List[int]: '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(_snake_case ) elif token_ids_a is None: return self._special_token_mask(_snake_case ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def UpperCAmelCase__ ( self : Any , _snake_case : Union[str, Any] , _snake_case : Tuple=None )->List[int]: '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def UpperCAmelCase__ ( self : Any , _snake_case : str , _snake_case : Optional[str] = None )->Tuple[str]: '''simple docstring''' if not os.path.isdir(_snake_case ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return __lowerCAmelCase : Optional[int] = os.path.join( _snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _snake_case ) elif not os.path.isfile(self.vocab_file ): with open(_snake_case , """wb""" ) as fi: __lowerCAmelCase : Tuple = self.sp_model.serialized_model_proto() fi.write(_snake_case ) return (out_vocab_file,)
232
1
import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( 'The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion' ) __lowerCAmelCase : str = None __lowerCAmelCase : Union[str, Any] = { '7B': 1_1008, '13B': 1_3824, '30B': 1_7920, '65B': 2_2016, '70B': 2_8672, } __lowerCAmelCase : Any = { '7B': 1, '7Bf': 1, '13B': 2, '13Bf': 2, '30B': 4, '65B': 8, '70B': 8, '70Bf': 8, } def __magic_name__ ( A : str, A : Union[str, Any]=1, A : Optional[int]=256 ): '''simple docstring''' return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def __magic_name__ ( A : List[Any] ): '''simple docstring''' with open(A, "r" ) as f: return json.load(A ) def __magic_name__ ( A : Tuple, A : Union[str, Any] ): '''simple docstring''' with open(A, "w" ) as f: json.dump(A, A ) def __magic_name__ ( A : str, A : Any, A : Optional[Any], A : str=True ): '''simple docstring''' os.makedirs(A, exist_ok=A ) a = os.path.join(A, "tmp" ) os.makedirs(A, exist_ok=A ) a = read_json(os.path.join(A, "params.json" ) ) a = NUM_SHARDS[model_size] a = params["n_layers"] a = params["n_heads"] a = n_heads // num_shards a = params["dim"] a = dim // n_heads a = 1_00_00.0 a = 1.0 / (base ** (torch.arange(0, A, 2 ).float() / dims_per_head)) if "n_kv_heads" in params: a = params["n_kv_heads"] # for GQA / MQA a = n_heads_per_shard // num_key_value_heads a = dim // num_key_value_heads else: # compatibility with other checkpoints a = n_heads a = n_heads_per_shard a = dim # permute for sliced rotary def permute(A : Any, A : Any=n_heads, A : Any=dim, A : Optional[int]=dim ): return w.view(A, dima // n_heads // 2, 2, A ).transpose(1, 2 ).reshape(A, A ) print(F"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) a = torch.load(os.path.join(A, "consolidated.00.pth" ), map_location="cpu" ) else: # Sharded a = [ torch.load(os.path.join(A, F"""consolidated.{i:02d}.pth""" ), map_location="cpu" ) for i in range(A ) ] a = 0 a = {"weight_map": {}} for layer_i in range(A ): a = F"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded a = { F"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[F"""layers.{layer_i}.attention.wq.weight"""] ), F"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[F"""layers.{layer_i}.attention.wk.weight"""] ), F"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[F"""layers.{layer_i}.attention.wv.weight"""], F"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[F"""layers.{layer_i}.attention.wo.weight"""], F"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[F"""layers.{layer_i}.feed_forward.w1.weight"""], F"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[F"""layers.{layer_i}.feed_forward.w2.weight"""], F"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[F"""layers.{layer_i}.feed_forward.w3.weight"""], F"""model.layers.{layer_i}.input_layernorm.weight""": loaded[F"""layers.{layer_i}.attention_norm.weight"""], F"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[F"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. a = { F"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ F"""layers.{layer_i}.attention_norm.weight""" ].clone(), F"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ F"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } a = permute( torch.cat( [ loaded[i][F"""layers.{layer_i}.attention.wq.weight"""].view(A, A, A ) for i in range(A ) ], dim=0, ).reshape(A, A ) ) a = permute( torch.cat( [ loaded[i][F"""layers.{layer_i}.attention.wk.weight"""].view( A, A, A ) for i in range(A ) ], dim=0, ).reshape(A, A ), A, A, A, ) a = torch.cat( [ loaded[i][F"""layers.{layer_i}.attention.wv.weight"""].view( A, A, A ) for i in range(A ) ], dim=0, ).reshape(A, A ) a = torch.cat( [loaded[i][F"""layers.{layer_i}.attention.wo.weight"""] for i in range(A )], dim=1 ) a = torch.cat( [loaded[i][F"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(A )], dim=0 ) a = torch.cat( [loaded[i][F"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(A )], dim=1 ) a = torch.cat( [loaded[i][F"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(A )], dim=0 ) a = inv_freq for k, v in state_dict.items(): a = filename param_count += v.numel() torch.save(A, os.path.join(A, A ) ) a = F"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded a = { "model.embed_tokens.weight": loaded["tok_embeddings.weight"], "model.norm.weight": loaded["norm.weight"], "lm_head.weight": loaded["output.weight"], } else: a = { "model.norm.weight": loaded[0]["norm.weight"], "model.embed_tokens.weight": torch.cat( [loaded[i]["tok_embeddings.weight"] for i in range(A )], dim=1 ), "lm_head.weight": torch.cat([loaded[i]["output.weight"] for i in range(A )], dim=0 ), } for k, v in state_dict.items(): a = filename param_count += v.numel() torch.save(A, os.path.join(A, A ) ) # Write configs a = {"total_size": param_count * 2} write_json(A, os.path.join(A, "pytorch_model.bin.index.json" ) ) a = params["ffn_dim_multiplier"] if "ffn_dim_multiplier" in params else 1 a = params["multiple_of"] if "multiple_of" in params else 256 a = LlamaConfig( hidden_size=A, intermediate_size=compute_intermediate_size(A, A, A ), num_attention_heads=params["n_heads"], num_hidden_layers=params["n_layers"], rms_norm_eps=params["norm_eps"], num_key_value_heads=A, ) config.save_pretrained(A ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print("Loading the checkpoint in a Llama model." ) a = LlamaForCausalLM.from_pretrained(A, torch_dtype=torch.floataa, low_cpu_mem_usage=A ) # Avoid saving this as part of the config. del model.config._name_or_path print("Saving in the Transformers format." ) model.save_pretrained(A, safe_serialization=A ) shutil.rmtree(A ) def __magic_name__ ( A : Optional[Any], A : Any ): '''simple docstring''' a = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(F"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) a = tokenizer_class(A ) tokenizer.save_pretrained(A ) def __magic_name__ ( ): '''simple docstring''' a = argparse.ArgumentParser() parser.add_argument( "--input_dir", help="Location of LLaMA weights, which contains tokenizer.model and model folders", ) parser.add_argument( "--model_size", choices=["7B", "7Bf", "13B", "13Bf", "30B", "65B", "70B", "70Bf", "tokenizer_only"], ) parser.add_argument( "--output_dir", help="Location to write HF model and tokenizer", ) parser.add_argument("--safe_serialization", type=A, help="Whether or not to save using `safetensors`." ) a = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir, input_base_path=os.path.join(args.input_dir, args.model_size ), model_size=args.model_size, safe_serialization=args.safe_serialization, ) a = os.path.join(args.input_dir, "tokenizer.model" ) write_tokenizer(args.output_dir, A ) if __name__ == "__main__": main()
107
from __future__ import annotations from scipy.special import comb # type: ignore class snake_case__ : """simple docstring""" def __init__( self : Any , __lowerCamelCase : list[tuple[float, float]] ) -> Tuple: a = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. a = len(__lowerCamelCase ) - 1 def __UpperCAmelCase ( self : Union[str, Any] , __lowerCamelCase : float ) -> list[float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." a = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __lowerCamelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__lowerCamelCase ) , 5 ) == 1 return output_values def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : float ) -> tuple[float, float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." a = self.basis_function(__lowerCamelCase ) a = 0.0 a = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : float = 0.01 ) -> List[str]: from matplotlib import pyplot as plt # type: ignore a = [] # x coordinates of points to plot a = [] # y coordinates of points to plot a = 0.0 while t <= 1: a = self.bezier_curve_function(__lowerCamelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size a = [i[0] for i in self.list_of_points] a = [i[1] for i in self.list_of_points] plt.plot( __lowerCamelCase , __lowerCamelCase , color="blue" , label="Curve of Degree " + str(self.degree ) , ) plt.scatter(__lowerCamelCase , __lowerCamelCase , color="red" , label="Control Points" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
107
1
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file A : str = 'Run commands across TPU VMs for initial setup before running `accelerate launch`.' def UpperCamelCase ( __magic_name__ : str=None ) -> Tuple: """simple docstring""" if subparsers is not None: lowercase__ = subparsers.add_parser("""tpu-config""" , description=_description ) else: lowercase__ = argparse.ArgumentParser("""Accelerate tpu-config command""" , description=_description ) # Core arguments lowercase__ = parser.add_argument_group( """Config Arguments""" , """Arguments that can be configured through `accelerate config`.""" ) config_args.add_argument( """--config_file""" , type=__magic_name__ , default=__magic_name__ , help="""Path to the config file to use for accelerate.""" , ) config_args.add_argument( """--tpu_name""" , default=__magic_name__ , help="""The name of the TPU to use. If not specified, will use the TPU specified in the config file.""" , ) config_args.add_argument( """--tpu_zone""" , default=__magic_name__ , help="""The zone of the TPU to use. If not specified, will use the zone specified in the config file.""" , ) lowercase__ = parser.add_argument_group("""TPU Arguments""" , """Arguments for options ran inside the TPU.""" ) pod_args.add_argument( """--use_alpha""" , action="""store_true""" , help="""Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.""" , ) pod_args.add_argument( """--command_file""" , default=__magic_name__ , help="""The path to the file containing the commands to run on the pod on startup.""" , ) pod_args.add_argument( """--command""" , action="""append""" , nargs="""+""" , help="""A command to run on the pod. Can be passed multiple times.""" , ) pod_args.add_argument( """--install_accelerate""" , action="""store_true""" , help="""Whether to install accelerate on the pod. Defaults to False.""" , ) pod_args.add_argument( """--accelerate_version""" , default="""latest""" , help="""The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub.""" , ) pod_args.add_argument( """--debug""" , action="""store_true""" , help="""If set, will print the command that would be run instead of running it.""" ) if subparsers is not None: parser.set_defaults(func=__magic_name__ ) return parser def UpperCamelCase ( __magic_name__ : int ) -> Optional[Any]: """simple docstring""" lowercase__ = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(__magic_name__ ): lowercase__ = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: lowercase__ = defaults.command_file if not args.command and defaults.commands is not None: lowercase__ = defaults.commands if not args.tpu_name: lowercase__ = defaults.tpu_name if not args.tpu_zone: lowercase__ = defaults.tpu_zone if args.accelerate_version == "dev": lowercase__ = """git+https://github.com/huggingface/accelerate.git""" elif args.accelerate_version == "latest": lowercase__ = """accelerate -U""" elif isinstance(parse(args.accelerate_version ) , __magic_name__ ): lowercase__ = f'''accelerate=={args.accelerate_version}''' if not args.command_file and not args.command: raise ValueError("""You must specify either a command file or a command to run on the pod.""" ) if args.command_file: with open(args.command_file , """r""" ) as f: lowercase__ = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , __magic_name__ ): lowercase__ = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate lowercase__ = ["""cd /usr/share"""] if args.install_accelerate: new_cmd += [f'''pip install {args.accelerate_version}'''] new_cmd += args.command lowercase__ = """; """.join(__magic_name__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess lowercase__ = ["""gcloud"""] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f'''Running {" ".join(__magic_name__ )}''' ) return subprocess.run(__magic_name__ ) print("""Successfully setup pod.""" ) def UpperCamelCase ( ) -> int: """simple docstring""" lowercase__ = tpu_command_parser() lowercase__ = parser.parse_args() tpu_command_launcher(__magic_name__ )
361
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class A ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase__ (self : Any ) -> Union[str, Any]: """simple docstring""" lowercase__ = TFXLMRobertaModel.from_pretrained("""jplu/tf-xlm-roberta-base""" ) lowercase__ = { """input_ids""": tf.convert_to_tensor([[0, 2646, 1_0269, 83, 9_9942, 2]] , dtype=tf.intaa ), # "My dog is cute" """attention_mask""": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } lowercase__ = model(_UpperCAmelCase )["""last_hidden_state"""] lowercase__ = tf.TensorShape((1, 6, 768) ) self.assertEqual(output.shape , _UpperCAmelCase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [ [ [0.0_681_762, 0.10_894_451, 0.06_772_504], [-0.06_423_668, 0.02_366_615, 0.04_329_344], [-0.06_057_295, 0.09_974_135, -0.00_070_584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
146
0
'''simple docstring''' import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class a__( unittest.TestCase ): '''simple docstring''' def a_ ( self): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights lowerCAmelCase = FlaxDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=__lowerCAmelCase , cache_dir=__lowerCAmelCase) lowerCAmelCase = [t[-1] for t in os.walk(os.path.join(__lowerCAmelCase , os.listdir(__lowerCAmelCase)[0] , """snapshots"""))] lowerCAmelCase = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(""".bin""") for f in files) @slow @require_flax class a__( unittest.TestCase ): '''simple docstring''' def a_ ( self): """simple docstring""" lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-pipe""" , safety_checker=__lowerCAmelCase) lowerCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) lowerCAmelCase = jax.random.PRNGKey(0) lowerCAmelCase = 4 lowerCAmelCase = jax.device_count() lowerCAmelCase = num_samples * [prompt] lowerCAmelCase = pipeline.prepare_inputs(__lowerCAmelCase) # shard inputs and rng lowerCAmelCase = replicate(__lowerCAmelCase) lowerCAmelCase = jax.random.split(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = shard(__lowerCAmelCase) lowerCAmelCase = pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa).sum() - 4.1514745) < 1E-3 assert np.abs(np.abs(__lowerCAmelCase , dtype=np.floataa).sum() - 49947.875) < 5E-1 lowerCAmelCase = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:]))) assert len(__lowerCAmelCase) == num_samples def a_ ( self): """simple docstring""" lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""flax""" , safety_checker=__lowerCAmelCase) lowerCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) lowerCAmelCase = jax.random.PRNGKey(0) lowerCAmelCase = 50 lowerCAmelCase = jax.device_count() lowerCAmelCase = num_samples * [prompt] lowerCAmelCase = pipeline.prepare_inputs(__lowerCAmelCase) # shard inputs and rng lowerCAmelCase = replicate(__lowerCAmelCase) lowerCAmelCase = jax.random.split(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = shard(__lowerCAmelCase) lowerCAmelCase = pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa).sum() - 0.05652401)) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa).sum() - 2383808.2)) < 5E-1 def a_ ( self): """simple docstring""" lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCAmelCase) lowerCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) lowerCAmelCase = jax.random.PRNGKey(0) lowerCAmelCase = 50 lowerCAmelCase = jax.device_count() lowerCAmelCase = num_samples * [prompt] lowerCAmelCase = pipeline.prepare_inputs(__lowerCAmelCase) # shard inputs and rng lowerCAmelCase = replicate(__lowerCAmelCase) lowerCAmelCase = jax.random.split(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = shard(__lowerCAmelCase) lowerCAmelCase = pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa).sum() - 0.04003906)) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa).sum() - 2373516.75)) < 5E-1 def a_ ( self): """simple docstring""" lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa) lowerCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) lowerCAmelCase = jax.random.PRNGKey(0) lowerCAmelCase = 50 lowerCAmelCase = jax.device_count() lowerCAmelCase = num_samples * [prompt] lowerCAmelCase = pipeline.prepare_inputs(__lowerCAmelCase) # shard inputs and rng lowerCAmelCase = replicate(__lowerCAmelCase) lowerCAmelCase = jax.random.split(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = shard(__lowerCAmelCase) lowerCAmelCase = pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa).sum() - 0.04003906)) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa).sum() - 2373516.75)) < 5E-1 def a_ ( self): """simple docstring""" lowerCAmelCase = FlaxDDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , set_alpha_to_one=__lowerCAmelCase , steps_offset=1 , ) lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , scheduler=__lowerCAmelCase , safety_checker=__lowerCAmelCase , ) lowerCAmelCase = scheduler.create_state() lowerCAmelCase = scheduler_state lowerCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) lowerCAmelCase = jax.random.PRNGKey(0) lowerCAmelCase = 50 lowerCAmelCase = jax.device_count() lowerCAmelCase = num_samples * [prompt] lowerCAmelCase = pipeline.prepare_inputs(__lowerCAmelCase) # shard inputs and rng lowerCAmelCase = replicate(__lowerCAmelCase) lowerCAmelCase = jax.random.split(__lowerCAmelCase , __lowerCAmelCase) lowerCAmelCase = shard(__lowerCAmelCase) lowerCAmelCase = pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa).sum() - 0.045043945)) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa).sum() - 2347693.5)) < 5E-1 def a_ ( self): """simple docstring""" lowerCAmelCase = ( """A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of""" """ field, close up, split lighting, cinematic""" ) lowerCAmelCase = jax.device_count() lowerCAmelCase = num_samples * [prompt] lowerCAmelCase = jax.random.split(jax.random.PRNGKey(0) , __lowerCAmelCase) lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCAmelCase , ) lowerCAmelCase = replicate(__lowerCAmelCase) lowerCAmelCase = pipeline.prepare_inputs(__lowerCAmelCase) lowerCAmelCase = shard(__lowerCAmelCase) lowerCAmelCase = pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase).images assert images.shape == (num_samples, 1, 512, 512, 3) lowerCAmelCase = images[2, 0, 256, 10:17, 1] # With memory efficient attention lowerCAmelCase = FlaxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""bf16""" , dtype=jnp.bfloataa , safety_checker=__lowerCAmelCase , use_memory_efficient_attention=__lowerCAmelCase , ) lowerCAmelCase = replicate(__lowerCAmelCase) lowerCAmelCase = pipeline.prepare_inputs(__lowerCAmelCase) lowerCAmelCase = shard(__lowerCAmelCase) lowerCAmelCase = pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) lowerCAmelCase = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice).max() < 1E-2
272
'''simple docstring''' from __future__ import annotations from typing import Any class __snake_case ( _SCREAMING_SNAKE_CASE): """simple docstring""" pass class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] , lowerCamelCase : Any ) -> None: lowerCAmelCase_ : Any = data lowerCAmelCase_ : Node | None = None def __iter__( self : Union[str, Any] ) -> Optional[Any]: lowerCAmelCase_ : Union[str, Any] = self lowerCAmelCase_ : Any = [] while node: if node in visited: raise ContainsLoopError visited.append(lowerCamelCase ) yield node.data lowerCAmelCase_ : int = node.next_node @property def __lowercase ( self : str ) -> bool: try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": __A : Dict = Node(1) __A : Optional[Any] = Node(2) __A : int = Node(3) __A : Optional[Any] = Node(4) print(root_node.has_loop) # False __A : Any = root_node.next_node print(root_node.has_loop) # True __A : List[Any] = Node(5) __A : Dict = Node(6) __A : str = Node(5) __A : Dict = Node(6) print(root_node.has_loop) # False __A : Optional[int] = Node(1) print(root_node.has_loop) # False
120
0
"""simple docstring""" import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class a ( unittest.TestCase ): def UpperCamelCase__ ( self ): """simple docstring""" super().tearDown() gc.collect() def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase ,lowerCAmelCase = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny' , from_pt=_snake_case , dtype=jnp.bfloataa ) lowerCAmelCase ,lowerCAmelCase = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=_snake_case , from_pt=_snake_case , dtype=jnp.bfloataa ) lowerCAmelCase = controlnet_params lowerCAmelCase = 'bird' lowerCAmelCase = jax.device_count() lowerCAmelCase = pipe.prepare_text_inputs([prompts] * num_samples ) lowerCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) lowerCAmelCase = pipe.prepare_image_inputs([canny_image] * num_samples ) lowerCAmelCase = jax.random.PRNGKey(0 ) lowerCAmelCase = jax.random.split(_snake_case , jax.device_count() ) lowerCAmelCase = replicate(_snake_case ) lowerCAmelCase = shard(_snake_case ) lowerCAmelCase = shard(_snake_case ) lowerCAmelCase = pipe( prompt_ids=_snake_case , image=_snake_case , params=_snake_case , prng_seed=_snake_case , num_inference_steps=50 , jit=_snake_case , ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) lowerCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCAmelCase = images[0, 2_53:2_56, 2_53:2_56, -1] lowerCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCAmelCase = jnp.array( [0.167_969, 0.116_699, 0.081_543, 0.154_297, 0.132_812, 0.108_887, 0.169_922, 0.169_922, 0.205_078] ) print(F'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase ,lowerCAmelCase = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose' , from_pt=_snake_case , dtype=jnp.bfloataa ) lowerCAmelCase ,lowerCAmelCase = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=_snake_case , from_pt=_snake_case , dtype=jnp.bfloataa ) lowerCAmelCase = controlnet_params lowerCAmelCase = 'Chef in the kitchen' lowerCAmelCase = jax.device_count() lowerCAmelCase = pipe.prepare_text_inputs([prompts] * num_samples ) lowerCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) lowerCAmelCase = pipe.prepare_image_inputs([pose_image] * num_samples ) lowerCAmelCase = jax.random.PRNGKey(0 ) lowerCAmelCase = jax.random.split(_snake_case , jax.device_count() ) lowerCAmelCase = replicate(_snake_case ) lowerCAmelCase = shard(_snake_case ) lowerCAmelCase = shard(_snake_case ) lowerCAmelCase = pipe( prompt_ids=_snake_case , image=_snake_case , params=_snake_case , prng_seed=_snake_case , num_inference_steps=50 , jit=_snake_case , ).images assert images.shape == (jax.device_count(), 1, 7_68, 5_12, 3) lowerCAmelCase = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCAmelCase = images[0, 2_53:2_56, 2_53:2_56, -1] lowerCAmelCase = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCAmelCase = jnp.array( [[0.271_484, 0.261_719, 0.275_391, 0.277_344, 0.279_297, 0.291_016, 0.294_922, 0.302_734, 0.302_734]] ) print(F'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
350
"""simple docstring""" from __future__ import annotations def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : list[str] | None = None ): lowerCAmelCase = word_bank or [] # create a table lowerCAmelCase = len(_UpperCAmelCase ) + 1 lowerCAmelCase = [] for _ in range(_UpperCAmelCase ): table.append([] ) # seed value lowerCAmelCase = [[]] # because empty string has empty combination # iterate through the indices for i in range(_UpperCAmelCase ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(_UpperCAmelCase )] == word: lowerCAmelCase = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(_UpperCAmelCase )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(_UpperCAmelCase )]: combination.reverse() return table[len(_UpperCAmelCase )] if __name__ == "__main__": print(all_construct('''jwajalapa''', ['''jwa''', '''j''', '''w''', '''a''', '''la''', '''lapa'''])) print(all_construct('''rajamati''', ['''s''', '''raj''', '''amat''', '''raja''', '''ma''', '''i''', '''t'''])) print( all_construct( '''hexagonosaurus''', ['''h''', '''ex''', '''hex''', '''ag''', '''ago''', '''ru''', '''auru''', '''rus''', '''go''', '''no''', '''o''', '''s'''], ) )
309
0
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging lowerCAmelCase : Tuple = logging.get_logger(__name__) lowerCAmelCase : Any = """▁""" lowerCAmelCase : List[Any] = { """vocab_file""": """vocab.json""", """spm_file""": """sentencepiece.bpe.model""", """tokenizer_config_file""": """tokenizer_config.json""", } lowerCAmelCase : List[str] = { """vocab_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json""", }, """spm_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model""", }, """tokenizer_config_file""": { """facebook/m2m100_418M""": """https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json""", """facebook/m2m100_1.2B""": """https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json""", }, } lowerCAmelCase : Optional[Any] = { """facebook/m2m100_418M""": 1024, } # fmt: off lowerCAmelCase : str = { """m2m100""": ["""af""", """am""", """ar""", """ast""", """az""", """ba""", """be""", """bg""", """bn""", """br""", """bs""", """ca""", """ceb""", """cs""", """cy""", """da""", """de""", """el""", """en""", """es""", """et""", """fa""", """ff""", """fi""", """fr""", """fy""", """ga""", """gd""", """gl""", """gu""", """ha""", """he""", """hi""", """hr""", """ht""", """hu""", """hy""", """id""", """ig""", """ilo""", """is""", """it""", """ja""", """jv""", """ka""", """kk""", """km""", """kn""", """ko""", """lb""", """lg""", """ln""", """lo""", """lt""", """lv""", """mg""", """mk""", """ml""", """mn""", """mr""", """ms""", """my""", """ne""", """nl""", """no""", """ns""", """oc""", """or""", """pa""", """pl""", """ps""", """pt""", """ro""", """ru""", """sd""", """si""", """sk""", """sl""", """so""", """sq""", """sr""", """ss""", """su""", """sv""", """sw""", """ta""", """th""", """tl""", """tn""", """tr""", """uk""", """ur""", """uz""", """vi""", """wo""", """xh""", """yi""", """yo""", """zh""", """zu"""], """wmt21""": ["""en""", """ha""", """is""", """ja""", """cs""", """ru""", """zh""", """de"""] } class __magic_name__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = ["input_ids", "attention_mask"] __UpperCamelCase = [] __UpperCamelCase = [] def __init__( self , _a , _a , _a=None , _a=None , _a="<s>" , _a="</s>" , _a="</s>" , _a="<pad>" , _a="<unk>" , _a="m2m100" , _a = None , _a=8 , **_a , ): """simple docstring""" lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs lowerCamelCase = language_codes lowerCamelCase = FAIRSEQ_LANGUAGE_CODES[language_codes] lowerCamelCase = {lang_code: f'__{lang_code}__' for lang_code in fairseq_language_code} lowerCamelCase = kwargs.get("""additional_special_tokens""" , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(_a ) for lang_code in fairseq_language_code if self.get_lang_token(_a ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=_a , tgt_lang=_a , bos_token=_a , eos_token=_a , sep_token=_a , unk_token=_a , pad_token=_a , language_codes=_a , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=_a , **_a , ) lowerCamelCase = vocab_file lowerCamelCase = load_json(_a ) lowerCamelCase = {v: k for k, v in self.encoder.items()} lowerCamelCase = spm_file lowerCamelCase = load_spm(_a , self.sp_model_kwargs ) lowerCamelCase = len(self.encoder ) lowerCamelCase = { self.get_lang_token(_a ): self.encoder_size + i for i, lang_code in enumerate(_a ) } lowerCamelCase = {lang_code: self.encoder_size + i for i, lang_code in enumerate(_a )} lowerCamelCase = {v: k for k, v in self.lang_token_to_id.items()} lowerCamelCase = src_lang if src_lang is not None else """en""" lowerCamelCase = tgt_lang lowerCamelCase = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) lowerCamelCase = num_madeup_words @property def _lowerCAmelCase ( self ): """simple docstring""" return len(self.encoder ) + len(self.lang_token_to_id ) @property def _lowerCAmelCase ( self ): """simple docstring""" return self._src_lang @src_lang.setter def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _lowerCAmelCase ( self , _a ): """simple docstring""" return self.sp_model.encode(_a , out_type=_a ) def _lowerCAmelCase ( self , _a ): """simple docstring""" if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(_a , self.encoder[self.unk_token] ) def _lowerCAmelCase ( self , _a ): """simple docstring""" if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(_a , self.unk_token ) def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = [] lowerCamelCase = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(_a ) + token lowerCamelCase = [] else: current_sub_tokens.append(_a ) out_string += self.sp_model.decode(_a ) return out_string.strip() def _lowerCAmelCase ( self , _a , _a = None , _a = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) lowerCamelCase = [1] * len(self.prefix_tokens ) lowerCamelCase = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(_a )) + suffix_ones return prefix_ones + ([0] * len(_a )) + ([0] * len(_a )) + suffix_ones def _lowerCAmelCase ( self , _a , _a = None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" lowerCamelCase = self.__dict__.copy() lowerCamelCase = None return state def __setstate__( self , _a ): """simple docstring""" lowerCamelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): lowerCamelCase = {} lowerCamelCase = load_spm(self.spm_file , self.sp_model_kwargs ) def _lowerCAmelCase ( self , _a , _a = None ): """simple docstring""" lowerCamelCase = Path(_a ) if not save_dir.is_dir(): raise OSError(f'{save_directory} should be a directory' ) lowerCamelCase = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""vocab_file"""] ) lowerCamelCase = save_dir / ( (filename_prefix + """-""" if filename_prefix else """""") + self.vocab_files_names["""spm_file"""] ) save_json(self.encoder , _a ) if os.path.abspath(self.spm_file ) != os.path.abspath(_a ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , _a ) elif not os.path.isfile(self.spm_file ): with open(_a , """wb""" ) as fi: lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(_a ) return (str(_a ), str(_a )) def _lowerCAmelCase ( self , _a , _a = "en" , _a = None , _a = "ro" , **_a , ): """simple docstring""" lowerCamelCase = src_lang lowerCamelCase = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(_a , _a , **_a ) def _lowerCAmelCase ( self , _a , _a , _a , **_a ): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) lowerCamelCase = src_lang lowerCamelCase = self(_a , add_special_tokens=_a , **_a ) lowerCamelCase = self.get_lang_id(_a ) lowerCamelCase = tgt_lang_id return inputs def _lowerCAmelCase ( self ): """simple docstring""" self.set_src_lang_special_tokens(self.src_lang ) def _lowerCAmelCase ( self ): """simple docstring""" self.set_tgt_lang_special_tokens(self.tgt_lang ) def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = self.get_lang_token(_a ) lowerCamelCase = self.lang_token_to_id[lang_token] lowerCamelCase = [self.cur_lang_id] lowerCamelCase = [self.eos_token_id] def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = self.get_lang_token(_a ) lowerCamelCase = self.lang_token_to_id[lang_token] lowerCamelCase = [self.cur_lang_id] lowerCamelCase = [self.eos_token_id] def _lowerCAmelCase ( self , _a ): """simple docstring""" return self.lang_code_to_token[lang] def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = self.get_lang_token(_a ) return self.lang_token_to_id[lang_token] def a__ ( snake_case__ , snake_case__ ) -> sentencepiece.SentencePieceProcessor: lowerCamelCase = sentencepiece.SentencePieceProcessor(**snake_case__ ) spm.Load(str(snake_case__ ) ) return spm def a__ ( snake_case__ ) -> Union[Dict, List]: with open(snake_case__ , """r""" ) as f: return json.load(snake_case__ ) def a__ ( snake_case__ , snake_case__ ) -> None: with open(snake_case__ , """w""" ) as f: json.dump(snake_case__ , snake_case__ , indent=2 )
291
"""simple docstring""" from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging lowerCAmelCase : Tuple = logging.get_logger(__name__) def a__ ( snake_case__ , snake_case__ ) -> Tuple: try: with open(snake_case__ , """rb""" ) as flax_state_f: lowerCamelCase = from_bytes(snake_case__ , flax_state_f.read() ) except UnpicklingError as e: try: with open(snake_case__ ) as f: if f.read().startswith("""version""" ): raise OSError( """You seem to have cloned a repository without having git-lfs installed. Please""" """ install git-lfs and run `git lfs install` followed by `git lfs pull` in the""" """ folder you cloned.""" ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(F'Unable to convert {model_file} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(snake_case__ , snake_case__ ) def a__ ( snake_case__ , snake_case__ ) -> Tuple: try: import torch # noqa: F401 except ImportError: logger.error( """Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights lowerCamelCase = flatten_dict(jax.tree_util.tree_map(lambda snake_case__ : x.dtype == jnp.bfloataa , snake_case__ ) ).values() if any(snake_case__ ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) lowerCamelCase = jax.tree_util.tree_map( lambda snake_case__ : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , snake_case__ ) lowerCamelCase = """""" lowerCamelCase = flatten_dict(snake_case__ , sep=""".""" ) lowerCamelCase = pt_model.state_dict() # keep track of unexpected & missing keys lowerCamelCase = [] lowerCamelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): lowerCamelCase = flax_key_tuple.split(""".""" ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: lowerCamelCase = flax_key_tuple_array[:-1] + ["""weight"""] lowerCamelCase = jnp.transpose(snake_case__ , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": lowerCamelCase = flax_key_tuple_array[:-1] + ["""weight"""] lowerCamelCase = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": lowerCamelCase = flax_key_tuple_array[:-1] + ["""weight"""] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(snake_case__ ): lowerCamelCase = ( flax_key_tuple_string.replace("""_0""" , """.0""" ) .replace("""_1""" , """.1""" ) .replace("""_2""" , """.2""" ) .replace("""_3""" , """.3""" ) .replace("""_4""" , """.4""" ) .replace("""_5""" , """.5""" ) .replace("""_6""" , """.6""" ) .replace("""_7""" , """.7""" ) .replace("""_8""" , """.8""" ) .replace("""_9""" , """.9""" ) ) lowerCamelCase = """.""".join(snake_case__ ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' F'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict lowerCamelCase = np.asarray(snake_case__ ) if not isinstance(snake_case__ , np.ndarray ) else flax_tensor lowerCamelCase = torch.from_numpy(snake_case__ ) # remove from missing keys missing_keys.remove(snake_case__ ) else: # weight is not expected by PyTorch model unexpected_keys.append(snake_case__ ) pt_model.load_state_dict(snake_case__ ) # re-transform missing_keys to list lowerCamelCase = list(snake_case__ ) if len(snake_case__ ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" F' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' F' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" F' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) if len(snake_case__ ) > 0: logger.warning( F'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' F' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' """ use it for predictions and inference.""" ) return pt_model
291
1
'''simple docstring''' import string def __lowerCamelCase ( lowerCAmelCase_ ) -> str: _a : List[Any] = '' for i in sequence: _a : Dict = ord(lowerCAmelCase_ ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extract <= 122: output += chr(219 - extract ) else: output += i return output def __lowerCamelCase ( lowerCAmelCase_ ) -> str: _a : Optional[int] = string.ascii_letters _a : str = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(lowerCAmelCase_ )] if c in letters else c for c in sequence ) def __lowerCamelCase ( ) -> None: from timeit import timeit print('Running performance benchmarks...' ) _a : Optional[Any] = 'from string import printable ; from __main__ import atbash, atbash_slow' print(f"""> atbash_slow(): {timeit('atbash_slow(printable)' , setup=lowerCAmelCase_ )} seconds""" ) print(f"""> atbash(): {timeit('atbash(printable)' , setup=lowerCAmelCase_ )} seconds""" ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(f"""{example} encrypted in atbash: {atbash(example)}""") benchmark()
107
'''simple docstring''' # This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES __lowerCAmelCase = '''tiny-wmt19-en-ru''' # Build # borrowed from a test __lowerCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] __lowerCAmelCase = dict(zip(vocab, range(len(vocab)))) __lowerCAmelCase = ['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = Path(tmpdirname) __lowerCAmelCase = build_dir / VOCAB_FILES_NAMES['''src_vocab_file'''] __lowerCAmelCase = build_dir / VOCAB_FILES_NAMES['''tgt_vocab_file'''] __lowerCAmelCase = build_dir / VOCAB_FILES_NAMES['''merges_file'''] with open(src_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, '''w''') as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, '''w''') as fp: fp.write('''\n'''.join(merges)) __lowerCAmelCase = FSMTTokenizer( langs=['''en''', '''ru'''], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) __lowerCAmelCase = FSMTConfig( langs=['''ru''', '''en'''], src_vocab_size=1_000, tgt_vocab_size=1_000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) __lowerCAmelCase = FSMTForConditionalGeneration(config) print(f"""num of params {tiny_model.num_parameters()}""") # Test __lowerCAmelCase = tokenizer(['''Making tiny model'''], return_tensors='''pt''') __lowerCAmelCase = tiny_model(**batch) print('''test output:''', len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(f"""Generated {mname_tiny}""") # Upload # transformers-cli upload tiny-wmt19-en-ru
107
1
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class a__ ( nn.Module ): _a : int _a : int _a : float = 0.0 _a : int = 1 _a : int = 1 _a : bool = True _a : bool = False _a : bool = False _a : bool = False _a : jnp.dtype = jnp.floataa def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = [] __lowerCAmelCase = [] for i in range(self.num_layers ): __lowerCAmelCase = self.in_channels if i == 0 else self.out_channels __lowerCAmelCase = FlaxResnetBlockaD( in_channels=SCREAMING_SNAKE_CASE__ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = resnets __lowerCAmelCase = attentions if self.add_downsample: __lowerCAmelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _A , _A , _A , _A=True ): """simple docstring""" __lowerCAmelCase = () for resnet, attn in zip(self.resnets , self.attentions ): __lowerCAmelCase = resnet(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = attn(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) output_states += (hidden_states,) if self.add_downsample: __lowerCAmelCase = self.downsamplers_a(SCREAMING_SNAKE_CASE__ ) output_states += (hidden_states,) return hidden_states, output_states class a__ ( nn.Module ): _a : int _a : int _a : float = 0.0 _a : int = 1 _a : bool = True _a : jnp.dtype = jnp.floataa def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = [] for i in range(self.num_layers ): __lowerCAmelCase = self.in_channels if i == 0 else self.out_channels __lowerCAmelCase = FlaxResnetBlockaD( in_channels=SCREAMING_SNAKE_CASE__ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = resnets if self.add_downsample: __lowerCAmelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _A , _A , _A=True ): """simple docstring""" __lowerCAmelCase = () for resnet in self.resnets: __lowerCAmelCase = resnet(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) output_states += (hidden_states,) if self.add_downsample: __lowerCAmelCase = self.downsamplers_a(SCREAMING_SNAKE_CASE__ ) output_states += (hidden_states,) return hidden_states, output_states class a__ ( nn.Module ): _a : int _a : int _a : int _a : float = 0.0 _a : int = 1 _a : int = 1 _a : bool = True _a : bool = False _a : bool = False _a : bool = False _a : jnp.dtype = jnp.floataa def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = [] __lowerCAmelCase = [] for i in range(self.num_layers ): __lowerCAmelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCAmelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCAmelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = resnets __lowerCAmelCase = attentions if self.add_upsample: __lowerCAmelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _A , _A , _A , _A , _A=True ): """simple docstring""" for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states __lowerCAmelCase = res_hidden_states_tuple[-1] __lowerCAmelCase = res_hidden_states_tuple[:-1] __lowerCAmelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCAmelCase = resnet(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = attn(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) if self.add_upsample: __lowerCAmelCase = self.upsamplers_a(SCREAMING_SNAKE_CASE__ ) return hidden_states class a__ ( nn.Module ): _a : int _a : int _a : int _a : float = 0.0 _a : int = 1 _a : bool = True _a : jnp.dtype = jnp.floataa def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = [] for i in range(self.num_layers ): __lowerCAmelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels __lowerCAmelCase = self.prev_output_channel if i == 0 else self.out_channels __lowerCAmelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = resnets if self.add_upsample: __lowerCAmelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , _A , _A , _A , _A=True ): """simple docstring""" for resnet in self.resnets: # pop res hidden states __lowerCAmelCase = res_hidden_states_tuple[-1] __lowerCAmelCase = res_hidden_states_tuple[:-1] __lowerCAmelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) __lowerCAmelCase = resnet(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) if self.add_upsample: __lowerCAmelCase = self.upsamplers_a(SCREAMING_SNAKE_CASE__ ) return hidden_states class a__ ( nn.Module ): _a : int _a : float = 0.0 _a : int = 1 _a : int = 1 _a : bool = False _a : bool = False _a : jnp.dtype = jnp.floataa def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] __lowerCAmelCase = [] for _ in range(self.num_layers ): __lowerCAmelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = resnets __lowerCAmelCase = attentions def __call__( self , _A , _A , _A , _A=True ): """simple docstring""" __lowerCAmelCase = self.resnets[0](SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): __lowerCAmelCase = attn(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = resnet(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , deterministic=SCREAMING_SNAKE_CASE__ ) return hidden_states
92
'''simple docstring''' from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging A_ = logging.get_logger(__name__) # pylint: disable=invalid-name class _snake_case ( _a ): def __init__( self : List[Any] ,SCREAMING_SNAKE_CASE__ : CLIPSegForImageSegmentation ,SCREAMING_SNAKE_CASE__ : CLIPSegProcessor ,SCREAMING_SNAKE_CASE__ : AutoencoderKL ,SCREAMING_SNAKE_CASE__ : CLIPTextModel ,SCREAMING_SNAKE_CASE__ : CLIPTokenizer ,SCREAMING_SNAKE_CASE__ : UNetaDConditionModel ,SCREAMING_SNAKE_CASE__ : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] ,SCREAMING_SNAKE_CASE__ : StableDiffusionSafetyChecker ,SCREAMING_SNAKE_CASE__ : CLIPImageProcessor ,): super().__init__() if hasattr(scheduler.config ,"steps_offset" ) and scheduler.config.steps_offset != 1: SCREAMING_SNAKE_CASE:Union[str, Any] = ( F'''The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`''' F''' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ''' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" ,"1.0.0" ,SCREAMING_SNAKE_CASE__ ,standard_warn=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE:Tuple = dict(scheduler.config ) SCREAMING_SNAKE_CASE:Union[str, Any] = 1 SCREAMING_SNAKE_CASE:Dict = FrozenDict(SCREAMING_SNAKE_CASE__ ) if hasattr(scheduler.config ,"skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: SCREAMING_SNAKE_CASE:List[Any] = ( F'''The configuration file of this scheduler: {scheduler} has not set the configuration''' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" ,"1.0.0" ,SCREAMING_SNAKE_CASE__ ,standard_warn=SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE:Tuple = dict(scheduler.config ) SCREAMING_SNAKE_CASE:int = True SCREAMING_SNAKE_CASE:Optional[int] = FrozenDict(SCREAMING_SNAKE_CASE__ ) if safety_checker is None: logger.warning( F'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=SCREAMING_SNAKE_CASE__ ,segmentation_processor=SCREAMING_SNAKE_CASE__ ,vae=SCREAMING_SNAKE_CASE__ ,text_encoder=SCREAMING_SNAKE_CASE__ ,tokenizer=SCREAMING_SNAKE_CASE__ ,unet=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ,safety_checker=SCREAMING_SNAKE_CASE__ ,feature_extractor=SCREAMING_SNAKE_CASE__ ,) def __UpperCamelCase ( self : int ,SCREAMING_SNAKE_CASE__ : Optional[Union[str, int]] = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory SCREAMING_SNAKE_CASE:Optional[Any] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(SCREAMING_SNAKE_CASE__ ) def __UpperCamelCase ( self : str ): self.enable_attention_slicing(SCREAMING_SNAKE_CASE__ ) def __UpperCamelCase ( self : List[str] ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) SCREAMING_SNAKE_CASE:str = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __UpperCamelCase ( self : Any ): if self.device != torch.device("meta" ) or not hasattr(self.unet ,"_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(SCREAMING_SNAKE_CASE__ ,"_hf_hook" ) and hasattr(module._hf_hook ,"execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : Dict ,SCREAMING_SNAKE_CASE__ : Union[str, List[str]] ,SCREAMING_SNAKE_CASE__ : Union[torch.FloatTensor, PIL.Image.Image] ,SCREAMING_SNAKE_CASE__ : str ,SCREAMING_SNAKE_CASE__ : int = 512 ,SCREAMING_SNAKE_CASE__ : int = 512 ,SCREAMING_SNAKE_CASE__ : int = 50 ,SCREAMING_SNAKE_CASE__ : float = 7.5 ,SCREAMING_SNAKE_CASE__ : Optional[Union[str, List[str]]] = None ,SCREAMING_SNAKE_CASE__ : Optional[int] = 1 ,SCREAMING_SNAKE_CASE__ : float = 0.0 ,SCREAMING_SNAKE_CASE__ : Optional[torch.Generator] = None ,SCREAMING_SNAKE_CASE__ : Optional[torch.FloatTensor] = None ,SCREAMING_SNAKE_CASE__ : Optional[str] = "pil" ,SCREAMING_SNAKE_CASE__ : bool = True ,SCREAMING_SNAKE_CASE__ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None ,SCREAMING_SNAKE_CASE__ : int = 1 ,**SCREAMING_SNAKE_CASE__ : Dict ,): SCREAMING_SNAKE_CASE:str = self.segmentation_processor( text=[text] ,images=[image] ,padding="max_length" ,return_tensors="pt" ).to(self.device ) SCREAMING_SNAKE_CASE:Union[str, Any] = self.segmentation_model(**SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE:Optional[int] = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() SCREAMING_SNAKE_CASE:Optional[Any] = self.numpy_to_pil(SCREAMING_SNAKE_CASE__ )[0].resize(image.size ) # Run inpainting pipeline with the generated mask SCREAMING_SNAKE_CASE:Any = StableDiffusionInpaintPipeline( vae=self.vae ,text_encoder=self.text_encoder ,tokenizer=self.tokenizer ,unet=self.unet ,scheduler=self.scheduler ,safety_checker=self.safety_checker ,feature_extractor=self.feature_extractor ,) return inpainting_pipeline( prompt=SCREAMING_SNAKE_CASE__ ,image=SCREAMING_SNAKE_CASE__ ,mask_image=SCREAMING_SNAKE_CASE__ ,height=SCREAMING_SNAKE_CASE__ ,width=SCREAMING_SNAKE_CASE__ ,num_inference_steps=SCREAMING_SNAKE_CASE__ ,guidance_scale=SCREAMING_SNAKE_CASE__ ,negative_prompt=SCREAMING_SNAKE_CASE__ ,num_images_per_prompt=SCREAMING_SNAKE_CASE__ ,eta=SCREAMING_SNAKE_CASE__ ,generator=SCREAMING_SNAKE_CASE__ ,latents=SCREAMING_SNAKE_CASE__ ,output_type=SCREAMING_SNAKE_CASE__ ,return_dict=SCREAMING_SNAKE_CASE__ ,callback=SCREAMING_SNAKE_CASE__ ,callback_steps=SCREAMING_SNAKE_CASE__ ,)
139
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase = {'''configuration_plbart''': ['''PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PLBartConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ['''PLBartTokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ '''PLBART_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PLBartForCausalLM''', '''PLBartForConditionalGeneration''', '''PLBartForSequenceClassification''', '''PLBartModel''', '''PLBartPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_plbart import PLBART_PRETRAINED_CONFIG_ARCHIVE_MAP, PLBartConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_plbart import PLBartTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_plbart import ( PLBART_PRETRAINED_MODEL_ARCHIVE_LIST, PLBartForCausalLM, PLBartForConditionalGeneration, PLBartForSequenceClassification, PLBartModel, PLBartPreTrainedModel, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
362
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { '''facebook/xlm-roberta-xl''': '''https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json''', '''facebook/xlm-roberta-xxl''': '''https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json''', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class _a ( _lowercase): _a : Optional[int] = '''xlm-roberta-xl''' def __init__( self : Any , _SCREAMING_SNAKE_CASE : str=25_0880 , _SCREAMING_SNAKE_CASE : Optional[Any]=2560 , _SCREAMING_SNAKE_CASE : int=36 , _SCREAMING_SNAKE_CASE : Optional[int]=32 , _SCREAMING_SNAKE_CASE : Any=1_0240 , _SCREAMING_SNAKE_CASE : List[str]="gelu" , _SCREAMING_SNAKE_CASE : Optional[Any]=0.1 , _SCREAMING_SNAKE_CASE : Dict=0.1 , _SCREAMING_SNAKE_CASE : Optional[int]=514 , _SCREAMING_SNAKE_CASE : Optional[int]=1 , _SCREAMING_SNAKE_CASE : Tuple=0.02 , _SCREAMING_SNAKE_CASE : Dict=1E-05 , _SCREAMING_SNAKE_CASE : Tuple=1 , _SCREAMING_SNAKE_CASE : Optional[int]=0 , _SCREAMING_SNAKE_CASE : Tuple=2 , _SCREAMING_SNAKE_CASE : Union[str, Any]="absolute" , _SCREAMING_SNAKE_CASE : Any=True , _SCREAMING_SNAKE_CASE : Any=None , **_SCREAMING_SNAKE_CASE : Tuple , )-> str: super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Any = vocab_size lowerCAmelCase__ : List[Any] = hidden_size lowerCAmelCase__ : int = num_hidden_layers lowerCAmelCase__ : Dict = num_attention_heads lowerCAmelCase__ : str = hidden_act lowerCAmelCase__ : List[Any] = intermediate_size lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : str = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : Any = type_vocab_size lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : Optional[int] = position_embedding_type lowerCAmelCase__ : Any = use_cache lowerCAmelCase__ : List[Any] = classifier_dropout class _a ( _lowercase): @property def UpperCAmelCase__( self : Any )-> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCAmelCase__ : int = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCAmelCase__ : List[Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
211
0
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("1.0.0a"): raise Exception("requires fairseq >= 1.0.0a") logging.set_verbosity_info() _lowerCamelCase : Any = logging.get_logger(__name__) _lowerCamelCase : Optional[Any] = "Hello world! cécé herlolip" def __lowerCamelCase ( A__ , A__ , A__ ) -> str: """simple docstring""" UpperCamelCase = FairseqRobertaModel.from_pretrained(A__ ) roberta.eval() # disable dropout UpperCamelCase = roberta.model.encoder.sentence_encoder UpperCamelCase = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings , hidden_size=roberta.cfg.model.encoder_embed_dim , num_hidden_layers=roberta.cfg.model.encoder_layers , num_attention_heads=roberta.cfg.model.encoder_attention_heads , intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1e-5 , ) if classification_head: UpperCamelCase = roberta.model.classification_heads['mnli'].out_proj.weight.shape[0] print('Our RoBERTa config:' , A__ ) UpperCamelCase = XLMRobertaXLForSequenceClassification(A__ ) if classification_head else XLMRobertaXLForMaskedLM(A__ ) model.eval() # Now let's copy all the weights. # Embeddings UpperCamelCase = roberta_sent_encoder.embed_tokens.weight UpperCamelCase = roberta_sent_encoder.embed_positions.weight UpperCamelCase = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. UpperCamelCase = roberta_sent_encoder.layer_norm.weight UpperCamelCase = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer UpperCamelCase = model.roberta.encoder.layer[i] UpperCamelCase = roberta_sent_encoder.layers[i] UpperCamelCase = layer.attention UpperCamelCase = roberta_layer.self_attn_layer_norm.weight UpperCamelCase = roberta_layer.self_attn_layer_norm.bias # self attention UpperCamelCase = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) UpperCamelCase = roberta_layer.self_attn.q_proj.weight UpperCamelCase = roberta_layer.self_attn.q_proj.bias UpperCamelCase = roberta_layer.self_attn.k_proj.weight UpperCamelCase = roberta_layer.self_attn.k_proj.bias UpperCamelCase = roberta_layer.self_attn.v_proj.weight UpperCamelCase = roberta_layer.self_attn.v_proj.bias # self-attention output UpperCamelCase = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape UpperCamelCase = roberta_layer.self_attn.out_proj.weight UpperCamelCase = roberta_layer.self_attn.out_proj.bias # this one is final layer norm UpperCamelCase = roberta_layer.final_layer_norm.weight UpperCamelCase = roberta_layer.final_layer_norm.bias # intermediate UpperCamelCase = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape UpperCamelCase = roberta_layer.fca.weight UpperCamelCase = roberta_layer.fca.bias # output UpperCamelCase = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape UpperCamelCase = roberta_layer.fca.weight UpperCamelCase = roberta_layer.fca.bias # end of layer if classification_head: UpperCamelCase = roberta.model.classification_heads['mnli'].dense.weight UpperCamelCase = roberta.model.classification_heads['mnli'].dense.bias UpperCamelCase = roberta.model.classification_heads['mnli'].out_proj.weight UpperCamelCase = roberta.model.classification_heads['mnli'].out_proj.bias else: # LM Head UpperCamelCase = roberta.model.encoder.lm_head.dense.weight UpperCamelCase = roberta.model.encoder.lm_head.dense.bias UpperCamelCase = roberta.model.encoder.lm_head.layer_norm.weight UpperCamelCase = roberta.model.encoder.lm_head.layer_norm.bias UpperCamelCase = roberta.model.encoder.lm_head.weight UpperCamelCase = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. UpperCamelCase = roberta.encode(A__ ).unsqueeze(0 ) # batch of size 1 UpperCamelCase = model(A__ )[0] if classification_head: UpperCamelCase = roberta.model.classification_heads['mnli'](roberta.extract_features(A__ ) ) else: UpperCamelCase = roberta.model(A__ )[0] print(our_output.shape , their_output.shape ) UpperCamelCase = torch.max(torch.abs(our_output - their_output ) ).item() print(F"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 UpperCamelCase = torch.allclose(A__ , A__ , atol=1e-3 ) print('Do both models output the same tensors?' , '🔥' if success else '💩' ) if not success: raise Exception('Something went wRoNg' ) pathlib.Path(A__ ).mkdir(parents=A__ , exist_ok=A__ ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(A__ ) if __name__ == "__main__": _lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--roberta_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--classification_head", action="store_true", help="Whether to convert a final classification head." ) _lowerCamelCase : Any = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
28
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor lowerCamelCase = logging.get_logger(__name__) class _a ( _lowercase): def __init__( self : Optional[int] , *_SCREAMING_SNAKE_CASE : Optional[Any] , **_SCREAMING_SNAKE_CASE : Optional[Any] )-> None: warnings.warn( '''The class PoolFormerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PoolFormerImageProcessor instead.''' , _SCREAMING_SNAKE_CASE , ) super().__init__(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )
131
0
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( "kwargs, expected", [ ({"num_shards": 0, "max_num_jobs": 1}, []), ({"num_shards": 10, "max_num_jobs": 1}, [range(10 )]), ({"num_shards": 10, "max_num_jobs": 10}, [range(__lowerCamelCase, i + 1 ) for i in range(10 )]), ({"num_shards": 1, "max_num_jobs": 10}, [range(1 )]), ({"num_shards": 10, "max_num_jobs": 3}, [range(0, 4 ), range(4, 7 ), range(7, 10 )]), ({"num_shards": 3, "max_num_jobs": 10}, [range(0, 1 ), range(1, 2 ), range(2, 3 )]), ], ) def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase ): _SCREAMING_SNAKE_CASE : int = _distribute_shards(**__lowerCamelCase ) assert out == expected @pytest.mark.parametrize( "gen_kwargs, max_num_jobs, expected", [ ({"foo": 0}, 10, [{"foo": 0}]), ({"shards": [0, 1, 2, 3]}, 1, [{"shards": [0, 1, 2, 3]}]), ({"shards": [0, 1, 2, 3]}, 4, [{"shards": [0]}, {"shards": [1]}, {"shards": [2]}, {"shards": [3]}]), ({"shards": [0, 1]}, 4, [{"shards": [0]}, {"shards": [1]}]), ({"shards": [0, 1, 2, 3]}, 2, [{"shards": [0, 1]}, {"shards": [2, 3]}]), ], ) def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): _SCREAMING_SNAKE_CASE : List[str] = _split_gen_kwargs(__lowerCamelCase, __lowerCamelCase ) assert out == expected @pytest.mark.parametrize( "gen_kwargs, expected", [ ({"foo": 0}, 1), ({"shards": [0]}, 1), ({"shards": [0, 1, 2, 3]}, 4), ({"shards": [0, 1, 2, 3], "foo": 0}, 4), ({"shards": [0, 1, 2, 3], "other": (0, 1)}, 4), ({"shards": [0, 1, 2, 3], "shards2": [0, 1]}, RuntimeError), ], ) def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase ): if expected is RuntimeError: with pytest.raises(__lowerCamelCase ): _number_of_shards_in_gen_kwargs(__lowerCamelCase ) else: _SCREAMING_SNAKE_CASE : Dict = _number_of_shards_in_gen_kwargs(__lowerCamelCase ) assert out == expected
351
import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever UpperCamelCase__ =logging.getLogger(__name__) class lowerCAmelCase__( __lowercase ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=None ) -> Optional[Any]: super().__init__( __lowerCamelCase , question_encoder_tokenizer=__lowerCamelCase , generator_tokenizer=__lowerCamelCase , index=__lowerCamelCase , init_retrieval=__lowerCamelCase , ) _SCREAMING_SNAKE_CASE : List[Any] = None def UpperCamelCase_ ( self , __lowerCamelCase ) -> Any: logger.info("initializing retrieval" ) # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("dist initialized" ) # needs to be set manually _SCREAMING_SNAKE_CASE : List[str] = self._infer_socket_ifname() # avoid clash with the NCCL port _SCREAMING_SNAKE_CASE : List[Any] = str(distributed_port + 1 ) _SCREAMING_SNAKE_CASE : int = dist.new_group(ranks=__lowerCamelCase , backend="gloo" ) # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("dist not initialized / main" ) self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group ) def UpperCamelCase_ ( self ) -> Optional[Any]: return dist.get_rank(group=self.process_group ) == 0 def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=torch.floataa ) -> Optional[Any]: _SCREAMING_SNAKE_CASE : Optional[int] = torch.empty(__lowerCamelCase , dtype=__lowerCamelCase ) dist.scatter(__lowerCamelCase , src=0 , scatter_list=__lowerCamelCase , group=self.process_group ) return target_tensor def UpperCamelCase_ ( self ) -> Tuple: _SCREAMING_SNAKE_CASE : int = psutil.net_if_addrs() # a hacky way to deal with varying network interface names _SCREAMING_SNAKE_CASE : Any = next((addr for addr in addrs if addr.startswith("e" )) , __lowerCamelCase ) return ifname def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase ) -> Tuple[np.ndarray, List[dict]]: # single GPU training if not dist.is_initialized(): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Dict = self._main_retrieve(__lowerCamelCase , __lowerCamelCase ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(__lowerCamelCase ) # distributed training _SCREAMING_SNAKE_CASE : Union[str, Any] = dist.get_world_size(group=self.process_group ) # gather logic _SCREAMING_SNAKE_CASE : Any = None if self._is_main(): _SCREAMING_SNAKE_CASE : Optional[Any] = [torch.empty(question_hidden_states.shape , dtype=torch.floataa ) for _ in range(__lowerCamelCase )] dist.gather(torch.tensor(__lowerCamelCase ) , dst=0 , gather_list=__lowerCamelCase , group=self.process_group ) # scatter logic _SCREAMING_SNAKE_CASE : Optional[int] = question_hidden_states.shape[0] _SCREAMING_SNAKE_CASE : Optional[Any] = [] _SCREAMING_SNAKE_CASE : Optional[int] = [] if self._is_main(): assert len(__lowerCamelCase ) == world_size _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : int = self._main_retrieve(torch.cat(__lowerCamelCase ).numpy() , __lowerCamelCase ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : List[str] = torch.tensor(__lowerCamelCase ), torch.tensor(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[str] = self._chunk_tensor(__lowerCamelCase , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Tuple = self._chunk_tensor(__lowerCamelCase , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = self._scattered(__lowerCamelCase , [n_queries, n_docs] , target_type=torch.intaa ) _SCREAMING_SNAKE_CASE : Optional[Any] = self._scattered(__lowerCamelCase , [n_queries, n_docs, question_hidden_states.shape[1]] ) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(__lowerCamelCase )
325
0
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
92
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class a__ ( snake_case__ ): def __init__( self , *_A , _A=None , _A=None , **_A ): """simple docstring""" super().__init__(*_A , **_A ) __lowerCAmelCase = eval_examples __lowerCAmelCase = post_process_function def __SCREAMING_SNAKE_CASE( self , _A = None , _A=None , _A = None , _A = "eval" , **_A , ): """simple docstring""" __lowerCAmelCase = gen_kwargs.copy() __lowerCAmelCase = ( gen_kwargs["max_length"] if gen_kwargs.get("max_length" ) is not None else self.args.generation_max_length ) __lowerCAmelCase = ( gen_kwargs["num_beams"] if gen_kwargs.get("num_beams" ) is not None else self.args.generation_num_beams ) __lowerCAmelCase = gen_kwargs __lowerCAmelCase = self.eval_dataset if eval_dataset is None else eval_dataset __lowerCAmelCase = self.get_eval_dataloader(_A ) __lowerCAmelCase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. __lowerCAmelCase = self.compute_metrics __lowerCAmelCase = None __lowerCAmelCase = time.time() __lowerCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: __lowerCAmelCase = eval_loop( _A , description="Evaluation" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_A , metric_key_prefix=_A , ) finally: __lowerCAmelCase = compute_metrics __lowerCAmelCase = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _A , _A , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default __lowerCAmelCase = self.post_process_function(_A , _A , _A ) __lowerCAmelCase = self.compute_metrics(_A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): __lowerCAmelCase = metrics.pop(_A ) metrics.update(output.metrics ) else: __lowerCAmelCase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(_A ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) __lowerCAmelCase = self.callback_handler.on_evaluate(self.args , self.state , self.control , _A ) return metrics def __SCREAMING_SNAKE_CASE( self , _A , _A , _A=None , _A = "test" , **_A ): """simple docstring""" __lowerCAmelCase = gen_kwargs.copy() __lowerCAmelCase = self.get_test_dataloader(_A ) # Temporarily disable metric computation, we will do it in the loop here. __lowerCAmelCase = self.compute_metrics __lowerCAmelCase = None __lowerCAmelCase = time.time() __lowerCAmelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: __lowerCAmelCase = eval_loop( _A , description="Prediction" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_A , metric_key_prefix=_A , ) finally: __lowerCAmelCase = compute_metrics __lowerCAmelCase = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _A , _A , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output __lowerCAmelCase = self.post_process_function(_A , _A , _A , "predict" ) __lowerCAmelCase = self.compute_metrics(_A ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): __lowerCAmelCase = metrics.pop(_A ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_A )
92
1
"""simple docstring""" from typing import Optional, Union import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models.modeling_utils import ModelMixin class _lowerCAmelCase ( a , a ): """simple docstring""" @register_to_config def __init__( self , __UpperCAmelCase = 7_6_8 , ): '''simple docstring''' super().__init__() lowerCAmelCase__ :str = nn.Parameter(torch.zeros(1 , __UpperCAmelCase ) ) lowerCAmelCase__ :Optional[Any] = nn.Parameter(torch.ones(1 , __UpperCAmelCase ) ) def snake_case ( self , __UpperCAmelCase = None , __UpperCAmelCase = None , ): '''simple docstring''' lowerCAmelCase__ :Dict = nn.Parameter(self.mean.to(__UpperCAmelCase ).to(__UpperCAmelCase ) ) lowerCAmelCase__ :Optional[Any] = nn.Parameter(self.std.to(__UpperCAmelCase ).to(__UpperCAmelCase ) ) return self def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :int = (embeds - self.mean) * 1.0 / self.std return embeds def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :str = (embeds * self.std) + self.mean return embeds
254
"""simple docstring""" from pathlib import Path import fire def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: """simple docstring""" lowerCAmelCase__ :List[str] = Path(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Any = Path(_SCREAMING_SNAKE_CASE ) dest_dir.mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) for path in src_dir.iterdir(): lowerCAmelCase__ :Union[str, Any] = [x.rstrip() for x in list(path.open().readlines() )][:n] lowerCAmelCase__ :Tuple = dest_dir.joinpath(path.name ) print(_SCREAMING_SNAKE_CASE ) dest_path.open('w' ).write('\n'.join(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": fire.Fire(minify)
254
1
"""simple docstring""" from __future__ import annotations from typing import Any class UpperCamelCase_ : """simple docstring""" def __init__( self : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : float = 0 ) -> None: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = row, column __SCREAMING_SNAKE_CASE = [[default_value for c in range(UpperCAmelCase__ )] for r in range(UpperCAmelCase__ )] def __str__( self : str ) -> str: __SCREAMING_SNAKE_CASE = F"""Matrix consist of {self.row} rows and {self.column} columns\n""" # Make string identifier __SCREAMING_SNAKE_CASE = 0 for row_vector in self.array: for obj in row_vector: __SCREAMING_SNAKE_CASE = max(UpperCAmelCase__ , len(str(UpperCAmelCase__ ) ) ) __SCREAMING_SNAKE_CASE = F"""%{max_element_length}s""" # Make string and return def single_line(UpperCAmelCase__ : list[float] ) -> str: nonlocal string_format_identifier __SCREAMING_SNAKE_CASE = "[" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(UpperCAmelCase__ ) for row_vector in self.array ) return s def __repr__( self : Union[str, Any] ) -> str: return str(self ) def UpperCAmelCase_ ( self : List[Any] , UpperCAmelCase__ : tuple[int, int] ) -> bool: if not (isinstance(UpperCAmelCase__ , (list, tuple) ) and len(UpperCAmelCase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple , UpperCAmelCase__ : tuple[int, int] ) -> Any: assert self.validate_indicies(UpperCAmelCase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Optional[int] , UpperCAmelCase__ : tuple[int, int] , UpperCAmelCase__ : float ) -> None: assert self.validate_indicies(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = value def __add__( self : Any , UpperCAmelCase__ : Matrix ) -> Matrix: assert isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) assert self.row == another.row and self.column == another.column # Add __SCREAMING_SNAKE_CASE = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __SCREAMING_SNAKE_CASE = self[r, c] + another[r, c] return result def __neg__( self : List[str] ) -> Matrix: __SCREAMING_SNAKE_CASE = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __SCREAMING_SNAKE_CASE = -self[r, c] return result def __sub__( self : Optional[int] , UpperCAmelCase__ : Matrix ) -> Matrix: return self + (-another) def __mul__( self : Optional[int] , UpperCAmelCase__ : int | float | Matrix ) -> Matrix: if isinstance(UpperCAmelCase__ , (int, float) ): # Scalar multiplication __SCREAMING_SNAKE_CASE = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __SCREAMING_SNAKE_CASE = self[r, c] * another return result elif isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): # Matrix multiplication assert self.column == another.row __SCREAMING_SNAKE_CASE = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __SCREAMING_SNAKE_CASE = F"""Unsupported type given for another ({type(UpperCAmelCase__ )})""" raise TypeError(UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Matrix: __SCREAMING_SNAKE_CASE = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): __SCREAMING_SNAKE_CASE = self[r, c] return result def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : Matrix , UpperCAmelCase__ : Matrix ) -> Any: assert isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) and isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __SCREAMING_SNAKE_CASE = v.transpose() __SCREAMING_SNAKE_CASE = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def UpperCAmelCase__ (): '''simple docstring''' __SCREAMING_SNAKE_CASE = Matrix(3 , 3 , 0 ) for i in range(3 ): __SCREAMING_SNAKE_CASE = 1 print(f"""a^(-1) is {ainv}""" ) # u, v __SCREAMING_SNAKE_CASE = Matrix(3 , 1 , 0 ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 1, 2, -3 __SCREAMING_SNAKE_CASE = Matrix(3 , 1 , 0 ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 4, -2, 5 print(f"""u is {u}""" ) print(f"""v is {v}""" ) print(f"""uv^T is {u * v.transpose()}""" ) # Sherman Morrison print(f"""(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCAmelCase_ , lowerCAmelCase_ )}""" ) def UpperCAmelCase__ (): '''simple docstring''' import doctest doctest.testmod() testa()
54
from __future__ import annotations def UpperCamelCase__ ( A__ , A__ , A__ ) -> tuple[float, list[float]]: snake_case__ : Optional[Any] = list(range(len(A__ ) ) ) snake_case__ : str = [v / w for v, w in zip(A__ , A__ )] index.sort(key=lambda A__ : ratio[i] , reverse=A__ ) snake_case__ : float = 0 snake_case__ : list[float] = [0] * len(A__ ) for i in index: if weight[i] <= capacity: snake_case__ : str = 1 max_value += value[i] capacity -= weight[i] else: snake_case__ : str = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
143
0
from __future__ import annotations from math import pi def lowerCamelCase__ ( lowercase , lowercase , lowercase ): """simple docstring""" if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if inductance < 0: raise ValueError("Inductance cannot be negative" ) if frequency < 0: raise ValueError("Frequency cannot be negative" ) if reactance < 0: raise ValueError("Inductive reactance cannot be negative" ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
357
def lowerCamelCase__ ( lowercase , lowercase = 0 ): """simple docstring""" SCREAMING_SNAKE_CASE : int = length or len(lowercase ) SCREAMING_SNAKE_CASE : Optional[Any] = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = list_data[i + 1], list_data[i] SCREAMING_SNAKE_CASE : str = True return list_data if not swapped else bubble_sort(lowercase , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
319
0
import requests from bsa import BeautifulSoup def UpperCamelCase ( lowerCAmelCase__ = "https://www.worldometers.info/coronavirus" ): '''simple docstring''' lowercase = BeautifulSoup(requests.get(lowerCAmelCase__ ).text , '''html.parser''' ) lowercase = soup.findAll('''h1''' ) lowercase = soup.findAll('''div''' , {'''class''': '''maincounter-number'''} ) keys += soup.findAll('''span''' , {'''class''': '''panel-title'''} ) values += soup.findAll('''div''' , {'''class''': '''number-table-main'''} ) return {key.text.strip(): value.text.strip() for key, value in zip(lowerCAmelCase__ , lowerCAmelCase__ )} if __name__ == "__main__": print("\033[1m" + "COVID-19 Status of the World" + "\033[0m\n") for key, value in world_covidaa_stats().items(): print(F'{key}\n{value}\n')
101
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import FEATURE_EXTRACTOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor, ChineseCLIPProcessor @require_vision class A_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self ): lowercase = tempfile.mkdtemp() lowercase = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', '的', '价', '格', '是', '15', '便', 'alex', '##andra', ',', '。', '-', 't', 'shirt', ] lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) lowercase = { 'do_resize': True, 'size': {'height': 224, 'width': 224}, 'do_center_crop': True, 'crop_size': {'height': 18, 'width': 18}, 'do_normalize': True, 'image_mean': [0.48_145_466, 0.4_578_275, 0.40_821_073], 'image_std': [0.26_862_954, 0.26_130_258, 0.27_577_711], 'do_convert_rgb': True, } lowercase = os.path.join(self.tmpdirname , snake_case ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , **snake_case ): return BertTokenizer.from_pretrained(self.tmpdirname , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , **snake_case ): return BertTokenizerFast.from_pretrained(self.tmpdirname , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , **snake_case ): return ChineseCLIPImageProcessor.from_pretrained(self.tmpdirname , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] lowercase = [Image.fromarray(np.moveaxis(snake_case , 0 , -1 ) ) for x in image_inputs] return image_inputs def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_tokenizer() lowercase = self.get_rust_tokenizer() lowercase = self.get_image_processor() lowercase = ChineseCLIPProcessor(tokenizer=snake_case , image_processor=snake_case ) processor_slow.save_pretrained(self.tmpdirname ) lowercase = ChineseCLIPProcessor.from_pretrained(self.tmpdirname , use_fast=snake_case ) lowercase = ChineseCLIPProcessor(tokenizer=snake_case , image_processor=snake_case ) processor_fast.save_pretrained(self.tmpdirname ) lowercase = ChineseCLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , snake_case ) self.assertIsInstance(processor_fast.tokenizer , snake_case ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , snake_case ) self.assertIsInstance(processor_fast.image_processor , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ChineseCLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) lowercase = self.get_tokenizer(cls_token='(CLS)' , sep_token='(SEP)' ) lowercase = self.get_image_processor(do_normalize=snake_case ) lowercase = ChineseCLIPProcessor.from_pretrained( self.tmpdirname , cls_token='(CLS)' , sep_token='(SEP)' , do_normalize=snake_case ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , snake_case ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = ChineseCLIPProcessor(tokenizer=snake_case , image_processor=snake_case ) lowercase = self.prepare_image_inputs() lowercase = image_processor(snake_case , return_tensors='np' ) lowercase = processor(images=snake_case , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = ChineseCLIPProcessor(tokenizer=snake_case , image_processor=snake_case ) lowercase = 'Alexandra,T-shirt的价格是15便士。' lowercase = processor(text=snake_case ) lowercase = tokenizer(snake_case ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = ChineseCLIPProcessor(tokenizer=snake_case , image_processor=snake_case ) lowercase = 'Alexandra,T-shirt的价格是15便士。' lowercase = self.prepare_image_inputs() lowercase = processor(text=snake_case , images=snake_case ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'token_type_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(snake_case ): processor() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = ChineseCLIPProcessor(tokenizer=snake_case , image_processor=snake_case ) lowercase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowercase = processor.batch_decode(snake_case ) lowercase = tokenizer.batch_decode(snake_case ) self.assertListEqual(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = ChineseCLIPProcessor(tokenizer=snake_case , image_processor=snake_case ) lowercase = 'Alexandra,T-shirt的价格是15便士。' lowercase = self.prepare_image_inputs() lowercase = processor(text=snake_case , images=snake_case ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
195
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def lowercase_ ( _lowerCamelCase : int=None) -> Any: if subparsers is not None: lowercase__ : Dict = subparsers.add_parser("test") else: lowercase__ : List[str] = argparse.ArgumentParser("Accelerate test command") parser.add_argument( "--config_file" , default=_lowerCamelCase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=_lowerCamelCase) return parser def lowercase_ ( _lowerCamelCase : List[Any]) -> Tuple: lowercase__ : List[str] = os.path.sep.join(__file__.split(os.path.sep)[:-2] + ["test_utils", "scripts", "test_script.py"]) if args.config_file is None: lowercase__ : Union[str, Any] = script_name else: lowercase__ : Union[str, Any] = f'''--config_file={args.config_file} {script_name}''' lowercase__ : List[str] = ["accelerate-launch"] + test_args.split() lowercase__ : Optional[Any] = execute_subprocess_async(_lowerCamelCase , env=os.environ.copy()) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!") def lowercase_ ( ) -> Optional[Any]: lowercase__ : int = test_command_parser() lowercase__ : Optional[int] = parser.parse_args() test_command(_lowerCamelCase) if __name__ == "__main__": main()
364
def lowercase_ ( _lowerCamelCase : int = 1 , _lowerCamelCase : int = 1000): lowercase__ : Union[str, Any] = 1 lowercase__ : int = 0 for divide_by_number in range(_lowerCamelCase , digit + 1): lowercase__ : list[int] = [] lowercase__ : Dict = numerator for _ in range(1 , digit + 1): if now_divide in has_been_divided: if longest_list_length < len(_lowerCamelCase): lowercase__ : Union[str, Any] = len(_lowerCamelCase) lowercase__ : Optional[int] = divide_by_number else: has_been_divided.append(_lowerCamelCase) lowercase__ : Optional[Any] = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
333
0
import os def _UpperCAmelCase ( a__): '''simple docstring''' a_ : Union[str, Any] = len(grid[0]) a_ : str = len(a__) a_ : Tuple = 0 a_ : List[Any] = 0 a_ : Optional[Any] = 0 # Check vertically, horizontally, diagonally at the same time (only works # for nxn grid) for i in range(a__): for j in range(n_rows - 3): a_ : Dict = grid[j][i] * grid[j + 1][i] * grid[j + 2][i] * grid[j + 3][i] a_ : Optional[int] = grid[i][j] * grid[i][j + 1] * grid[i][j + 2] * grid[i][j + 3] # Left-to-right diagonal (\) product if i < n_columns - 3: a_ : List[str] = ( grid[i][j] * grid[i + 1][j + 1] * grid[i + 2][j + 2] * grid[i + 3][j + 3] ) # Right-to-left diagonal(/) product if i > 2: a_ : Optional[int] = ( grid[i][j] * grid[i - 1][j + 1] * grid[i - 2][j + 2] * grid[i - 3][j + 3] ) a_ : Optional[int] = max( a__ , a__ , a__ , a__) if max_product > largest: a_ : Union[str, Any] = max_product return largest def _UpperCAmelCase ( ): '''simple docstring''' a_ : Dict = [] with open(os.path.dirname(a__) + """/grid.txt""") as file: for line in file: grid.append(line.strip("""\n""").split(""" """)) a_ : int = [[int(a__) for i in grid[j]] for j in range(len(a__))] return largest_product(a__) if __name__ == "__main__": print(solution())
248
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() __snake_case : Union[str, Any] = logging.get_logger(__name__) def _UpperCAmelCase ( a__ , a__ , a__ , a__): '''simple docstring''' a_ : Tuple = original_name.split(""".""")[0] a_ : List[Any] = key.split(""".""") a_ : List[Any] = int(key_list[key_list.index(a__) - 2]) a_ : Dict = int(key_list[key_list.index(a__) - 1]) a_ : Any = orig_block_num - offset a_ : Optional[int] = key.replace(f'''{orig_block_num}.{layer_num}.{original_name}''' , f'''block.{new_block_num}.{layer_num}.{new_name}''') return key def _UpperCAmelCase ( a__): '''simple docstring''' a_ : List[str] = OrderedDict() a_ , a_ : Optional[int] = 0, 0 for key, value in state_dict.items(): if key.startswith("""network"""): a_ : str = key.replace("""network""" , """poolformer.encoder""") if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("""bias""") and "patch_embed" not in key: patch_emb_offset += 1 a_ : Tuple = key[: key.find("""proj""")] a_ : Dict = key.replace(a__ , f'''patch_embeddings.{total_embed_found}.''') a_ : Optional[Any] = key.replace("""proj""" , """projection""") if key.endswith("""bias"""): total_embed_found += 1 if "patch_embeddings" in key: a_ : int = """poolformer.encoder.""" + key if "mlp.fc1" in key: a_ : Union[str, Any] = replace_key_with_offset(a__ , a__ , """mlp.fc1""" , """output.conv1""") if "mlp.fc2" in key: a_ : str = replace_key_with_offset(a__ , a__ , """mlp.fc2""" , """output.conv2""") if "norm1" in key: a_ : str = replace_key_with_offset(a__ , a__ , """norm1""" , """before_norm""") if "norm2" in key: a_ : Any = replace_key_with_offset(a__ , a__ , """norm2""" , """after_norm""") if "layer_scale_1" in key: a_ : List[Any] = replace_key_with_offset(a__ , a__ , """layer_scale_1""" , """layer_scale_1""") if "layer_scale_2" in key: a_ : Optional[Any] = replace_key_with_offset(a__ , a__ , """layer_scale_2""" , """layer_scale_2""") if "head" in key: a_ : Optional[Any] = key.replace("""head""" , """classifier""") a_ : Union[str, Any] = value return new_state_dict def _UpperCAmelCase ( ): '''simple docstring''' a_ : Optional[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" a_ : Any = Image.open(requests.get(a__ , stream=a__).raw) return image @torch.no_grad() def _UpperCAmelCase ( a__ , a__ , a__): '''simple docstring''' a_ : str = PoolFormerConfig() # set attributes based on model_name a_ : Union[str, Any] = """huggingface/label-files""" a_ : str = model_name[-3:] a_ : Tuple = 1_0_0_0 a_ : List[str] = """imagenet-1k-id2label.json""" a_ : Any = (1, 1_0_0_0) # set config attributes a_ : Optional[Any] = json.load(open(hf_hub_download(a__ , a__ , repo_type="""dataset""") , """r""")) a_ : List[Any] = {int(a__): v for k, v in idalabel.items()} a_ : Tuple = idalabel a_ : int = {v: k for k, v in idalabel.items()} if size == "s12": a_ : Optional[int] = [2, 2, 6, 2] a_ : str = [6_4, 1_2_8, 3_2_0, 5_1_2] a_ : List[Any] = 4.0 a_ : Tuple = 0.9 elif size == "s24": a_ : List[Any] = [4, 4, 1_2, 4] a_ : str = [6_4, 1_2_8, 3_2_0, 5_1_2] a_ : List[Any] = 4.0 a_ : Optional[Any] = 0.9 elif size == "s36": a_ : str = [6, 6, 1_8, 6] a_ : Dict = [6_4, 1_2_8, 3_2_0, 5_1_2] a_ : Optional[int] = 4.0 a_ : Optional[int] = 1e-6 a_ : Tuple = 0.9 elif size == "m36": a_ : str = [6, 6, 1_8, 6] a_ : List[Any] = [9_6, 1_9_2, 3_8_4, 7_6_8] a_ : str = 4.0 a_ : Union[str, Any] = 1e-6 a_ : str = 0.95 elif size == "m48": a_ : List[Any] = [8, 8, 2_4, 8] a_ : Dict = [9_6, 1_9_2, 3_8_4, 7_6_8] a_ : int = 4.0 a_ : int = 1e-6 a_ : List[Any] = 0.95 else: raise ValueError(f'''Size {size} not supported''') # load image processor a_ : Tuple = PoolFormerImageProcessor(crop_pct=a__) # Prepare image a_ : List[Any] = prepare_img() a_ : List[str] = image_processor(images=a__ , return_tensors="""pt""").pixel_values logger.info(f'''Converting model {model_name}...''') # load original state dict a_ : List[str] = torch.load(a__ , map_location=torch.device("""cpu""")) # rename keys a_ : List[Any] = rename_keys(a__) # create HuggingFace model and load state dict a_ : List[str] = PoolFormerForImageClassification(a__) model.load_state_dict(a__) model.eval() # Define image processor a_ : Tuple = PoolFormerImageProcessor(crop_pct=a__) a_ : Any = image_processor(images=prepare_img() , return_tensors="""pt""").pixel_values # forward pass a_ : Any = model(a__) a_ : Any = outputs.logits # define expected logit slices for different models if size == "s12": a_ : Union[str, Any] = torch.tensor([-0.3045, -0.6758, -0.4869]) elif size == "s24": a_ : Optional[Any] = torch.tensor([0.4402, -0.1374, -0.8045]) elif size == "s36": a_ : int = torch.tensor([-0.6080, -0.5133, -0.5898]) elif size == "m36": a_ : List[str] = torch.tensor([0.3952, 0.2263, -1.2668]) elif size == "m48": a_ : Union[str, Any] = torch.tensor([0.1167, -0.0656, -0.3423]) else: raise ValueError(f'''Size {size} not supported''') # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , a__ , atol=1e-2) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''') Path(a__).mkdir(exist_ok=a__) model.save_pretrained(a__) print(f'''Saving image processor to {pytorch_dump_folder_path}''') image_processor.save_pretrained(a__) if __name__ == "__main__": __snake_case : List[str] = argparse.ArgumentParser() parser.add_argument( """--model_name""", default="""poolformer_s12""", type=str, help="""Name of the model you'd like to convert.""", ) parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) __snake_case : Optional[int] = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
248
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType A_ = logging.get_logger(__name__) A_ = { '''openai/imagegpt-small''': '''''', '''openai/imagegpt-medium''': '''''', '''openai/imagegpt-large''': '''''', } class lowercase( __a ): '''simple docstring''' lowercase__ = "imagegpt" lowercase__ = ["past_key_values"] lowercase__ = { "hidden_size": "n_embd", "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self: Union[str, Any], a_: Union[str, Any]=512 + 1, a_: int=32 * 32, a_: str=512, a_: List[Any]=24, a_: Any=8, a_: int=None, a_: Optional[int]="quick_gelu", a_: Optional[Any]=0.1, a_: Any=0.1, a_: Any=0.1, a_: Optional[Any]=1E-5, a_: List[str]=0.02, a_: int=True, a_: Union[str, Any]=True, a_: int=False, a_: Dict=False, a_: Tuple=False, **a_: Any, ): '''simple docstring''' _snake_case : List[str] = vocab_size _snake_case : Optional[int] = n_positions _snake_case : str = n_embd _snake_case : Any = n_layer _snake_case : Optional[Any] = n_head _snake_case : Tuple = n_inner _snake_case : Optional[int] = activation_function _snake_case : int = resid_pdrop _snake_case : Dict = embd_pdrop _snake_case : List[str] = attn_pdrop _snake_case : List[Any] = layer_norm_epsilon _snake_case : List[Any] = initializer_range _snake_case : Union[str, Any] = scale_attn_weights _snake_case : Optional[Any] = use_cache _snake_case : int = scale_attn_by_inverse_layer_idx _snake_case : int = reorder_and_upcast_attn _snake_case : Dict = tie_word_embeddings super().__init__(tie_word_embeddings=a_, **a_ ) class lowercase( __a ): '''simple docstring''' @property def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ] ) def UpperCamelCase_ ( self: List[Any], a_: "FeatureExtractionMixin", a_: int = 1, a_: int = -1, a_: bool = False, a_: Optional["TensorType"] = None, a_: int = 3, a_: int = 32, a_: int = 32, ): '''simple docstring''' _snake_case : Tuple = self._generate_dummy_images(a_, a_, a_, a_ ) _snake_case : List[str] = dict(preprocessor(images=a_, return_tensors=a_ ) ) return inputs
366
"""simple docstring""" from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
132
0
from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
232
class lowerCamelCase__ : '''simple docstring''' def __init__( self :Union[str, Any] ) -> None: __UpperCamelCase : dict[str, TrieNode] = {} # Mapping from char to TrieNode __UpperCamelCase : List[str] = False def _lowerCamelCase ( self :Any , a :list[str] ) -> None: for word in words: self.insert(a ) def _lowerCamelCase ( self :List[str] , a :str ) -> None: __UpperCamelCase : Dict = self for char in word: if char not in curr.nodes: __UpperCamelCase : List[Any] = TrieNode() __UpperCamelCase : List[Any] = curr.nodes[char] __UpperCamelCase : Union[str, Any] = True def _lowerCamelCase ( self :Optional[int] , a :str ) -> bool: __UpperCamelCase : Union[str, Any] = self for char in word: if char not in curr.nodes: return False __UpperCamelCase : Union[str, Any] = curr.nodes[char] return curr.is_leaf def _lowerCamelCase ( self :Any , a :str ) -> None: def _delete(a :TrieNode , a :str , a :int ) -> bool: if index == len(a ): # If word does not exist if not curr.is_leaf: return False __UpperCamelCase : str = False return len(curr.nodes ) == 0 __UpperCamelCase : List[Any] = word[index] __UpperCamelCase : Optional[int] = curr.nodes.get(a ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted __UpperCamelCase : int = _delete(a , a , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , a , 0 ) def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : TrieNode , _lowerCamelCase : str) -> None: '''simple docstring''' if node.is_leaf: print(_lowerCamelCase , end=" ") for key, value in node.nodes.items(): print_words(_lowerCamelCase , word + key) def _SCREAMING_SNAKE_CASE ( ) -> bool: '''simple docstring''' __UpperCamelCase : int = "banana bananas bandana band apple all beast".split() __UpperCamelCase : Union[str, Any] = TrieNode() root.insert_many(_lowerCamelCase) # print_words(root, "") assert all(root.find(_lowerCamelCase) for word in words) assert root.find("banana") assert not root.find("bandanas") assert not root.find("apps") assert root.find("apple") assert root.find("all") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : str , _lowerCamelCase : bool) -> None: '''simple docstring''' print(str(_lowerCamelCase) , "works!" if passes else "doesn't work :(") def _SCREAMING_SNAKE_CASE ( ) -> None: '''simple docstring''' assert test_trie() def _SCREAMING_SNAKE_CASE ( ) -> None: '''simple docstring''' print_results("Testing trie functionality" , test_trie()) if __name__ == "__main__": main()
232
1
'''simple docstring''' import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: lowercase__ : str = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class SCREAMING_SNAKE_CASE (unittest.TestCase ): def __init__( self , _UpperCAmelCase , _UpperCAmelCase=7 , _UpperCAmelCase=3 , _UpperCAmelCase=18 , _UpperCAmelCase=30 , _UpperCAmelCase=400 , _UpperCAmelCase=None , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=None , ): '''simple docstring''' __A : List[str] = size if size is not None else {'height': 20, 'width': 20} __A : Any = parent __A : Optional[int] = batch_size __A : List[Any] = num_channels __A : Optional[Any] = image_size __A : int = min_resolution __A : str = max_resolution __A : Optional[int] = size __A : int = do_normalize __A : Tuple = do_convert_rgb __A : List[str] = [512, 1024, 2048, 4096] __A : str = patch_size if patch_size is not None else {'height': 16, 'width': 16} def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Union[str, Any] = 'https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg' __A : str = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase).raw).convert('RGB') return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class SCREAMING_SNAKE_CASE (a__ , unittest.TestCase ): lowerCAmelCase = PixaStructImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Any = PixaStructImageProcessingTester(self) @property def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : int = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(_UpperCAmelCase , 'do_normalize')) self.assertTrue(hasattr(_UpperCAmelCase , 'do_convert_rgb')) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[str] = self.image_processor_tester.prepare_dummy_image() __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict) __A : List[str] = 2048 __A : Any = image_processor(_UpperCAmelCase , return_tensors='pt' , max_patches=_UpperCAmelCase) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.0606) , atol=1e-3 , rtol=1e-3)) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Any = self.image_processing_class(**self.image_processor_dict) # create random PIL images __A : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , Image.Image) # Test not batched input __A : int = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __A : Dict = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __A : str = image_processor( _UpperCAmelCase , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[str] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __A : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , Image.Image) # Test not batched input __A : Optional[int] = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 __A : List[str] = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(_UpperCAmelCase): __A : Union[str, Any] = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches __A : int = 'Hello' __A : Dict = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=_UpperCAmelCase , header_text=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __A : Tuple = image_processor( _UpperCAmelCase , return_tensors='pt' , max_patches=_UpperCAmelCase , header_text=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __A : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase , numpify=_UpperCAmelCase) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , np.ndarray) __A : str = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __A : Optional[int] = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __A : Union[str, Any] = image_processor( _UpperCAmelCase , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __A : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase , torchify=_UpperCAmelCase) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , torch.Tensor) # Test not batched input __A : str = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __A : Optional[Any] = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __A : List[Any] = image_processor( _UpperCAmelCase , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class SCREAMING_SNAKE_CASE (a__ , unittest.TestCase ): lowerCAmelCase = PixaStructImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = PixaStructImageProcessingTester(self , num_channels=4) __A : Optional[int] = 3 @property def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(_UpperCAmelCase , 'do_normalize')) self.assertTrue(hasattr(_UpperCAmelCase , 'do_convert_rgb')) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Tuple = self.image_processing_class(**self.image_processor_dict) # create random PIL images __A : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_UpperCAmelCase) for image in image_inputs: self.assertIsInstance(_UpperCAmelCase , Image.Image) # Test not batched input __A : Any = ( (self.image_processor_tester.patch_size['height'] * self.image_processor_tester.patch_size['width']) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input __A : Optional[int] = image_processor( image_inputs[0] , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched __A : Any = image_processor( _UpperCAmelCase , return_tensors='pt' , max_patches=_UpperCAmelCase).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
190
'''simple docstring''' from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration lowercase__ : List[Any] = HfArgumentParser(InitializationArguments) lowercase__ : List[Any] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization lowercase__ : Union[str, Any] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks lowercase__ : List[str] = { '''vocab_size''': len(tokenizer), '''scale_attn_by_inverse_layer_idx''': True, '''reorder_and_upcast_attn''': True, } # Load model config (GPT-2 large in this case) lowercase__ : Optional[int] = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config lowercase__ : Optional[int] = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
190
1
from __future__ import annotations import math def a__ ( A_ ): '''simple docstring''' if num <= 0: __magic_name__ = f'''{num}: Invalid input, please enter a positive integer.''' raise ValueError(A_ ) __magic_name__ = [True] * (num + 1) __magic_name__ = [] __magic_name__ = 2 __magic_name__ = int(math.sqrt(A_ ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(A_ ) # Set multiples of start be False for i in range(start * start, num + 1, A_ ): if sieve[i] is True: __magic_name__ = False start += 1 for j in range(end + 1, num + 1 ): if sieve[j] is True: prime.append(A_ ) return prime if __name__ == "__main__": print(prime_sieve(int(input('Enter a positive integer: ').strip())))
88
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class a_ : """simple docstring""" def __init__( self : Optional[int] ,snake_case : Any ,snake_case : Dict=100 ,snake_case : List[Any]=13 ,snake_case : str=30 ,snake_case : List[str]=2 ,snake_case : List[Any]=3 ,snake_case : Tuple=True ,snake_case : Optional[Any]=True ,snake_case : int=32 ,snake_case : Tuple=4 ,snake_case : List[Any]=4 ,snake_case : Optional[Any]=37 ,snake_case : Optional[Any]="gelu" ,snake_case : Tuple=0.1 ,snake_case : Union[str, Any]=0.1 ,snake_case : List[Any]=10 ,snake_case : Tuple=0.02 ,snake_case : List[str]=3 ,snake_case : Any=None ,snake_case : int=[0, 1, 2, 3] ,): SCREAMING_SNAKE_CASE =parent SCREAMING_SNAKE_CASE =100 SCREAMING_SNAKE_CASE =batch_size SCREAMING_SNAKE_CASE =image_size SCREAMING_SNAKE_CASE =patch_size SCREAMING_SNAKE_CASE =num_channels SCREAMING_SNAKE_CASE =is_training SCREAMING_SNAKE_CASE =use_labels SCREAMING_SNAKE_CASE =hidden_size SCREAMING_SNAKE_CASE =num_hidden_layers SCREAMING_SNAKE_CASE =num_attention_heads SCREAMING_SNAKE_CASE =intermediate_size SCREAMING_SNAKE_CASE =hidden_act SCREAMING_SNAKE_CASE =hidden_dropout_prob SCREAMING_SNAKE_CASE =attention_probs_dropout_prob SCREAMING_SNAKE_CASE =type_sequence_label_size SCREAMING_SNAKE_CASE =initializer_range SCREAMING_SNAKE_CASE =scope SCREAMING_SNAKE_CASE =out_indices SCREAMING_SNAKE_CASE =num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) SCREAMING_SNAKE_CASE =(image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE =num_patches + 1 def _lowerCAmelCase ( self : List[Any] ): SCREAMING_SNAKE_CASE =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE =None SCREAMING_SNAKE_CASE =None if self.use_labels: SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size] ,self.type_sequence_label_size ) SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.image_size, self.image_size] ,self.num_labels ) SCREAMING_SNAKE_CASE =self.get_config() return config, pixel_values, labels, pixel_labels def _lowerCAmelCase ( self : Dict ): return BeitConfig( vocab_size=self.vocab_size ,image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=snake_case ,initializer_range=self.initializer_range ,out_indices=self.out_indices ,) def _lowerCAmelCase ( self : Union[str, Any] ,snake_case : Tuple ,snake_case : Optional[Any] ,snake_case : Union[str, Any] ,snake_case : Optional[int] ): SCREAMING_SNAKE_CASE =BeitModel(config=snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCAmelCase ( self : Union[str, Any] ,snake_case : Optional[int] ,snake_case : Dict ,snake_case : Any ,snake_case : List[str] ): SCREAMING_SNAKE_CASE =BeitForMaskedImageModeling(config=snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length - 1, self.vocab_size) ) def _lowerCAmelCase ( self : Optional[Any] ,snake_case : Any ,snake_case : str ,snake_case : Any ,snake_case : str ): SCREAMING_SNAKE_CASE =self.type_sequence_label_size SCREAMING_SNAKE_CASE =BeitForImageClassification(snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ,labels=snake_case ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE =1 SCREAMING_SNAKE_CASE =BeitForImageClassification(snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE =model(snake_case ,labels=snake_case ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def _lowerCAmelCase ( self : List[str] ,snake_case : Tuple ,snake_case : str ,snake_case : Optional[int] ,snake_case : int ): SCREAMING_SNAKE_CASE =self.num_labels SCREAMING_SNAKE_CASE =BeitForSemanticSegmentation(snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) SCREAMING_SNAKE_CASE =model(snake_case ,labels=snake_case ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def _lowerCAmelCase ( self : str ): SCREAMING_SNAKE_CASE =self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =config_and_inputs SCREAMING_SNAKE_CASE ={'pixel_values': pixel_values} return config, inputs_dict @require_torch class a_ ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase = ( { 'feature-extraction': BeitModel, 'image-classification': BeitForImageClassification, 'image-segmentation': BeitForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = False def _lowerCAmelCase ( self : List[Any] ): SCREAMING_SNAKE_CASE =BeitModelTester(self ) SCREAMING_SNAKE_CASE =ConfigTester(self ,config_class=snake_case ,has_text_modality=snake_case ,hidden_size=37 ) def _lowerCAmelCase ( self : List[str] ): self.config_tester.run_common_tests() @unittest.skip(reason='BEiT does not use inputs_embeds' ) def _lowerCAmelCase ( self : List[Any] ): pass @require_torch_multi_gpu @unittest.skip(reason='BEiT has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def _lowerCAmelCase ( self : Union[str, Any] ): pass def _lowerCAmelCase ( self : Tuple ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE =model_class(snake_case ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) SCREAMING_SNAKE_CASE =model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case ,nn.Linear ) ) def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE =model_class(snake_case ) SCREAMING_SNAKE_CASE =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE =[*signature.parameters.keys()] SCREAMING_SNAKE_CASE =['pixel_values'] self.assertListEqual(arg_names[:1] ,snake_case ) def _lowerCAmelCase ( self : List[str] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _lowerCAmelCase ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def _lowerCAmelCase ( self : Dict ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case ) def _lowerCAmelCase ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*snake_case ) def _lowerCAmelCase ( self : Any ): if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE =True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(snake_case ), BeitForMaskedImageModeling]: continue SCREAMING_SNAKE_CASE =model_class(snake_case ) model.to(snake_case ) model.train() SCREAMING_SNAKE_CASE =self._prepare_for_class(snake_case ,snake_case ,return_labels=snake_case ) SCREAMING_SNAKE_CASE =model(**snake_case ).loss loss.backward() def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE =False SCREAMING_SNAKE_CASE =True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(snake_case ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue SCREAMING_SNAKE_CASE =model_class(snake_case ) model.gradient_checkpointing_enable() model.to(snake_case ) model.train() SCREAMING_SNAKE_CASE =self._prepare_for_class(snake_case ,snake_case ,return_labels=snake_case ) SCREAMING_SNAKE_CASE =model(**snake_case ).loss loss.backward() def _lowerCAmelCase ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE =_config_zero_init(snake_case ) for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE =model_class(config=snake_case ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() ,[0.0, 1.0] ,msg=f'Parameter {name} of model {model_class} seems not properly initialized' ,) @slow def _lowerCAmelCase ( self : List[str] ): for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE =BeitModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def snake_case__ ( ): """simple docstring""" SCREAMING_SNAKE_CASE =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class a_ ( unittest.TestCase ): """simple docstring""" @cached_property def _lowerCAmelCase ( self : Tuple ): return BeitImageProcessor.from_pretrained('microsoft/beit-base-patch16-224' ) if is_vision_available() else None @slow def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =BeitForMaskedImageModeling.from_pretrained('microsoft/beit-base-patch16-224-pt22k' ).to(snake_case ) SCREAMING_SNAKE_CASE =self.default_image_processor SCREAMING_SNAKE_CASE =prepare_img() SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).pixel_values.to(snake_case ) # prepare bool_masked_pos SCREAMING_SNAKE_CASE =torch.ones((1, 196) ,dtype=torch.bool ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(pixel_values=snake_case ,bool_masked_pos=snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 196, 8192) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =torch.tensor( [[-3.2_437, 0.5_072, -13.9_174], [-3.2_456, 0.4_948, -13.9_401], [-3.2_033, 0.5_121, -13.8_550]] ).to(snake_case ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] ,snake_case ,atol=1e-2 ) ) @slow def _lowerCAmelCase ( self : List[str] ): SCREAMING_SNAKE_CASE =BeitForImageClassification.from_pretrained('microsoft/beit-base-patch16-224' ).to(snake_case ) SCREAMING_SNAKE_CASE =self.default_image_processor SCREAMING_SNAKE_CASE =prepare_img() SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 1000) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =torch.tensor([-1.2_385, -1.0_987, -1.0_108] ).to(snake_case ) self.assertTrue(torch.allclose(logits[0, :3] ,snake_case ,atol=1e-4 ) ) SCREAMING_SNAKE_CASE =281 self.assertEqual(logits.argmax(-1 ).item() ,snake_case ) @slow def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =BeitForImageClassification.from_pretrained('microsoft/beit-large-patch16-224-pt22k-ft22k' ).to( snake_case ) SCREAMING_SNAKE_CASE =self.default_image_processor SCREAMING_SNAKE_CASE =prepare_img() SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 21841) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =torch.tensor([1.6_881, -0.2_787, 0.5_901] ).to(snake_case ) self.assertTrue(torch.allclose(logits[0, :3] ,snake_case ,atol=1e-4 ) ) SCREAMING_SNAKE_CASE =2396 self.assertEqual(logits.argmax(-1 ).item() ,snake_case ) @slow def _lowerCAmelCase ( self : Tuple ): SCREAMING_SNAKE_CASE =BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) SCREAMING_SNAKE_CASE =model.to(snake_case ) SCREAMING_SNAKE_CASE =BeitImageProcessor(do_resize=snake_case ,size=640 ,do_center_crop=snake_case ) SCREAMING_SNAKE_CASE =load_dataset('hf-internal-testing/fixtures_ade20k' ,split='test' ) SCREAMING_SNAKE_CASE =Image.open(ds[0]['file'] ) SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =version.parse(PIL.__version__ ) < version.parse('9.0.0' ) if is_pillow_less_than_a: SCREAMING_SNAKE_CASE =torch.tensor( [ [[-4.9_225, -2.3_954, -3.0_522], [-2.8_822, -1.0_046, -1.7_561], [-2.9_549, -1.3_228, -2.1_347]], [[-5.8_168, -3.4_129, -4.0_778], [-3.8_651, -2.2_214, -3.0_277], [-3.8_356, -2.4_643, -3.3_535]], [[-0.0_078, 3.9_952, 4.0_754], [2.9_856, 4.6_944, 5.0_035], [3.2_413, 4.7_813, 4.9_969]], ] ,device=snake_case ,) else: SCREAMING_SNAKE_CASE =torch.tensor( [ [[-4.8_960, -2.3_688, -3.0_355], [-2.8_478, -0.9_836, -1.7_418], [-2.9_449, -1.3_332, -2.1_456]], [[-5.8_081, -3.4_124, -4.1_006], [-3.8_561, -2.2_081, -3.0_323], [-3.8_365, -2.4_601, -3.3_669]], [[-0.0_309, 3.9_868, 4.0_540], [2.9_640, 4.6_877, 4.9_976], [3.2_081, 4.7_690, 4.9_942]], ] ,device=snake_case ,) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] ,snake_case ,atol=1e-4 ) ) @slow def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) SCREAMING_SNAKE_CASE =model.to(snake_case ) SCREAMING_SNAKE_CASE =BeitImageProcessor(do_resize=snake_case ,size=640 ,do_center_crop=snake_case ) SCREAMING_SNAKE_CASE =load_dataset('hf-internal-testing/fixtures_ade20k' ,split='test' ) SCREAMING_SNAKE_CASE =Image.open(ds[0]['file'] ) SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits.detach().cpu() SCREAMING_SNAKE_CASE =image_processor.post_process_semantic_segmentation(outputs=snake_case ,target_sizes=[(500, 300)] ) SCREAMING_SNAKE_CASE =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape ,snake_case ) SCREAMING_SNAKE_CASE =image_processor.post_process_semantic_segmentation(outputs=snake_case ) SCREAMING_SNAKE_CASE =torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape ,snake_case )
334
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A = { """configuration_nllb_moe""": [ """NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NllbMoeConfig""", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ """NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST""", """NllbMoeForConditionalGeneration""", """NllbMoeModel""", """NllbMoePreTrainedModel""", """NllbMoeTop2Router""", """NllbMoeSparseMLP""", ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys _A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
354
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _lowerCamelCase ( a_ ): _lowerCamelCase :Dict = ["image_processor", "tokenizer"] _lowerCamelCase :Dict = "BlipImageProcessor" _lowerCamelCase :Any = ("BertTokenizer", "BertTokenizerFast") def __init__( self : Dict , UpperCamelCase : Optional[int] , UpperCamelCase : Union[str, Any] ) -> str: """simple docstring""" lowerCAmelCase__ : Optional[Any] = False super().__init__(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : List[Any] = self.image_processor def __call__( self : int , UpperCamelCase : ImageInput = None , UpperCamelCase : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , UpperCamelCase : bool = True , UpperCamelCase : Union[bool, str, PaddingStrategy] = False , UpperCamelCase : Union[bool, str, TruncationStrategy] = None , UpperCamelCase : Optional[int] = None , UpperCamelCase : int = 0 , UpperCamelCase : Optional[int] = None , UpperCamelCase : Optional[bool] = None , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = False , UpperCamelCase : bool = True , UpperCamelCase : Optional[Union[str, TensorType]] = None , **UpperCamelCase : Optional[int] , ) -> BatchEncoding: """simple docstring""" if images is None and text is None: raise ValueError("""You have to specify either images or text.""" ) # Get only text if images is None: lowerCAmelCase__ : Any = self.tokenizer lowerCAmelCase__ : Optional[int] = self.tokenizer( text=UpperCamelCase , add_special_tokens=UpperCamelCase , padding=UpperCamelCase , truncation=UpperCamelCase , max_length=UpperCamelCase , stride=UpperCamelCase , pad_to_multiple_of=UpperCamelCase , return_attention_mask=UpperCamelCase , return_overflowing_tokens=UpperCamelCase , return_special_tokens_mask=UpperCamelCase , return_offsets_mapping=UpperCamelCase , return_token_type_ids=UpperCamelCase , return_length=UpperCamelCase , verbose=UpperCamelCase , return_tensors=UpperCamelCase , **UpperCamelCase , ) return text_encoding # add pixel_values lowerCAmelCase__ : Tuple = self.image_processor(UpperCamelCase , return_tensors=UpperCamelCase ) if text is not None: lowerCAmelCase__ : Optional[int] = self.tokenizer( text=UpperCamelCase , add_special_tokens=UpperCamelCase , padding=UpperCamelCase , truncation=UpperCamelCase , max_length=UpperCamelCase , stride=UpperCamelCase , pad_to_multiple_of=UpperCamelCase , return_attention_mask=UpperCamelCase , return_overflowing_tokens=UpperCamelCase , return_special_tokens_mask=UpperCamelCase , return_offsets_mapping=UpperCamelCase , return_token_type_ids=UpperCamelCase , return_length=UpperCamelCase , verbose=UpperCamelCase , return_tensors=UpperCamelCase , **UpperCamelCase , ) else: lowerCAmelCase__ : Tuple = None if text_encoding is not None: encoding_image_processor.update(UpperCamelCase ) return encoding_image_processor def _lowerCAmelCase ( self : int , *UpperCamelCase : Union[str, Any] , **UpperCamelCase : List[str] ) -> Dict: """simple docstring""" return self.tokenizer.batch_decode(*UpperCamelCase , **UpperCamelCase ) def _lowerCAmelCase ( self : List[Any] , *UpperCamelCase : Tuple , **UpperCamelCase : List[str] ) -> List[Any]: """simple docstring""" return self.tokenizer.decode(*UpperCamelCase , **UpperCamelCase ) @property def _lowerCAmelCase ( self : List[Any] ) -> Tuple: """simple docstring""" lowerCAmelCase__ : Optional[int] = self.tokenizer.model_input_names lowerCAmelCase__ : Union[str, Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
212
0
"""simple docstring""" from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging __A = logging.get_logger(__name__) __A = { "EleutherAI/gpt-j-6B": "https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :Dict = "gptj" _UpperCAmelCase :List[Any] = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , _UpperCAmelCase=50400 , _UpperCAmelCase=2048 , _UpperCAmelCase=4096 , _UpperCAmelCase=28 , _UpperCAmelCase=16 , _UpperCAmelCase=64 , _UpperCAmelCase=None , _UpperCAmelCase="gelu_new" , _UpperCAmelCase=0.0 , _UpperCAmelCase=0.0 , _UpperCAmelCase=0.0 , _UpperCAmelCase=1e-5 , _UpperCAmelCase=0.02 , _UpperCAmelCase=True , _UpperCAmelCase=50256 , _UpperCAmelCase=50256 , _UpperCAmelCase=False , **_UpperCAmelCase , ): lowercase__: Optional[int] = vocab_size lowercase__: int = n_positions lowercase__: Optional[Any] = n_embd lowercase__: Dict = n_layer lowercase__: str = n_head lowercase__: int = n_inner lowercase__: Tuple = rotary_dim lowercase__: Any = activation_function lowercase__: Optional[Any] = resid_pdrop lowercase__: str = embd_pdrop lowercase__: Union[str, Any] = attn_pdrop lowercase__: str = layer_norm_epsilon lowercase__: str = initializer_range lowercase__: str = use_cache lowercase__: Optional[int] = bos_token_id lowercase__: str = eos_token_id super().__init__( bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , tie_word_embeddings=_UpperCAmelCase , **_UpperCAmelCase ) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase = "default" , _UpperCAmelCase = None , _UpperCAmelCase = False , ): super().__init__(_UpperCAmelCase , task=_UpperCAmelCase , patching_specs=_UpperCAmelCase , use_past=_UpperCAmelCase ) if not getattr(self._config , '''pad_token_id''' , _UpperCAmelCase ): # TODO: how to do that better? lowercase__: Any = 0 @property def _snake_case ( self ): lowercase__: int = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}} ) if self.use_past: self.fill_with_past_key_values_(_UpperCAmelCase , direction='''inputs''' ) lowercase__: List[Any] = {0: '''batch''', 1: '''past_sequence + sequence'''} else: lowercase__: Any = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def _snake_case ( self ): return self._config.n_layer @property def _snake_case ( self ): return self._config.n_head def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase = -1 , _UpperCAmelCase = -1 , _UpperCAmelCase = False , _UpperCAmelCase = None , ): lowercase__: List[Any] = super(_UpperCAmelCase , self ).generate_dummy_inputs( _UpperCAmelCase , batch_size=_UpperCAmelCase , seq_length=_UpperCAmelCase , is_pair=_UpperCAmelCase , framework=_UpperCAmelCase ) # We need to order the input in the way they appears in the forward() lowercase__: Optional[Any] = OrderedDict({'''input_ids''': common_inputs['''input_ids''']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase__, lowercase__: Tuple = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase__: Any = seqlen + 2 lowercase__: Any = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowercase__: Dict = [ (torch.zeros(_UpperCAmelCase ), torch.zeros(_UpperCAmelCase )) for _ in range(self.num_layers ) ] lowercase__: List[str] = common_inputs['''attention_mask'''] if self.use_past: lowercase__: Union[str, Any] = ordered_inputs['''attention_mask'''].dtype lowercase__: List[Any] = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(_UpperCAmelCase , _UpperCAmelCase , dtype=_UpperCAmelCase )] , dim=1 ) return ordered_inputs @property def _snake_case ( self ): return 13
177
"""simple docstring""" def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase=2_8_1_2_3 ) -> Any: lowercase__: Optional[Any] = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i lowercase__: Union[str, Any] = set() lowercase__: Optional[Any] = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(__UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
177
1
'''simple docstring''' from __future__ import annotations def a_ ( lowerCamelCase : int ): lowerCAmelCase = 2 lowerCAmelCase = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(lowerCamelCase ) if n > 1: factors.append(lowerCamelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
55
'''simple docstring''' import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class UpperCAmelCase_ ( __lowercase , __lowercase , __lowercase , unittest.TestCase ): lowerCamelCase : Any = StableUnCLIPPipeline lowerCamelCase : int = TEXT_TO_IMAGE_PARAMS lowerCamelCase : Any = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase : Optional[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false lowerCamelCase : Optional[int] = False def __UpperCAmelCase ( self : Optional[Any] ) -> Tuple: lowerCAmelCase = 3_2 lowerCAmelCase = embedder_hidden_size # prior components torch.manual_seed(0 ) lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) torch.manual_seed(0 ) lowerCAmelCase = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=UpperCAmelCase__ , projection_dim=UpperCAmelCase__ , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) ) torch.manual_seed(0 ) lowerCAmelCase = PriorTransformer( num_attention_heads=2 , attention_head_dim=1_2 , embedding_dim=UpperCAmelCase__ , num_layers=1 , ) torch.manual_seed(0 ) lowerCAmelCase = DDPMScheduler( variance_type='fixed_small_log' , prediction_type='sample' , num_train_timesteps=1_0_0_0 , clip_sample=UpperCAmelCase__ , clip_sample_range=5.0 , beta_schedule='squaredcos_cap_v2' , ) # regular denoising components torch.manual_seed(0 ) lowerCAmelCase = StableUnCLIPImageNormalizer(embedding_dim=UpperCAmelCase__ ) lowerCAmelCase = DDPMScheduler(beta_schedule='squaredcos_cap_v2' ) torch.manual_seed(0 ) lowerCAmelCase = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) torch.manual_seed(0 ) lowerCAmelCase = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=UpperCAmelCase__ , projection_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) ) torch.manual_seed(0 ) lowerCAmelCase = UNetaDConditionModel( sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock2D', 'DownBlock2D') , up_block_types=('UpBlock2D', 'CrossAttnUpBlock2D') , block_out_channels=(3_2, 6_4) , attention_head_dim=(2, 4) , class_embed_type='projection' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=UpperCAmelCase__ , layers_per_block=1 , upcast_attention=UpperCAmelCase__ , use_linear_projection=UpperCAmelCase__ , ) torch.manual_seed(0 ) lowerCAmelCase = DDIMScheduler( beta_schedule='scaled_linear' , beta_start=0.00_085 , beta_end=0.012 , prediction_type='v_prediction' , set_alpha_to_one=UpperCAmelCase__ , steps_offset=1 , ) torch.manual_seed(0 ) lowerCAmelCase = AutoencoderKL() lowerCAmelCase = { # prior components 'prior_tokenizer': prior_tokenizer, 'prior_text_encoder': prior_text_encoder, 'prior': prior, 'prior_scheduler': prior_scheduler, # image noising components 'image_normalizer': image_normalizer, 'image_noising_scheduler': image_noising_scheduler, # regular denoising components 'tokenizer': tokenizer, 'text_encoder': text_encoder, 'unet': unet, 'scheduler': scheduler, 'vae': vae, } return components def __UpperCAmelCase ( self : Any , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any]=0 ) -> Optional[Any]: if str(UpperCAmelCase__ ).startswith('mps' ): lowerCAmelCase = torch.manual_seed(UpperCAmelCase__ ) else: lowerCAmelCase = torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) lowerCAmelCase = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'prior_num_inference_steps': 2, 'output_type': 'numpy', } return inputs def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: lowerCAmelCase = torch_device == 'cpu' self._test_attention_slicing_forward_pass(test_max_difference=UpperCAmelCase__ ) def __UpperCAmelCase ( self : int ) -> Union[str, Any]: lowerCAmelCase = torch_device in ['cpu', 'mps'] self._test_inference_batch_single_identical(test_max_difference=UpperCAmelCase__ ) @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self : Union[str, Any] ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Tuple ) -> Union[str, Any]: lowerCAmelCase = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy' ) lowerCAmelCase = StableUnCLIPPipeline.from_pretrained('fusing/stable-unclip-2-1-l' , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCAmelCase = torch.Generator(device='cpu' ).manual_seed(0 ) lowerCAmelCase = pipe('anime turle' , generator=UpperCAmelCase__ , output_type='np' ) lowerCAmelCase = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(UpperCAmelCase__ , UpperCAmelCase__ ) def __UpperCAmelCase ( self : Any ) -> Optional[int]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCAmelCase = StableUnCLIPPipeline.from_pretrained('fusing/stable-unclip-2-1-l' , torch_dtype=torch.floataa ) lowerCAmelCase = pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCAmelCase = pipe( 'anime turtle' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='np' , ) lowerCAmelCase = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 1_0**9
55
1
from typing import List from .keymap import KEYMAP, get_character def lowerCamelCase__ ( _a): def decorator(_a): SCREAMING_SNAKE_CASE : Dict = getattr(_a , "handle_key" , []) handle += [key] setattr(_a , "handle_key" , _a) return func return decorator def lowerCamelCase__ ( *_a): def decorator(_a): SCREAMING_SNAKE_CASE : Dict = getattr(_a , "handle_key" , []) handle += keys setattr(_a , "handle_key" , _a) return func return decorator class _UpperCamelCase ( __A ): '''simple docstring''' def __new__( cls : Tuple , a : Tuple , a : Dict , a : int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = super().__new__(cls , a , a , a ) if not hasattr(a , "key_handler" ): setattr(a , "key_handler" , {} ) setattr(a , "handle_input" , KeyHandler.handle_input ) for value in attrs.values(): SCREAMING_SNAKE_CASE : List[Any] = getattr(a , "handle_key" , [] ) for key in handled_keys: SCREAMING_SNAKE_CASE : List[str] = value return new_cls @staticmethod def __UpperCamelCase ( cls : Any ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = get_character() if char != KEYMAP["undefined"]: SCREAMING_SNAKE_CASE : int = ord(a ) SCREAMING_SNAKE_CASE : str = cls.key_handler.get(a ) if handler: SCREAMING_SNAKE_CASE : Dict = char return handler(cls ) else: return None def lowerCamelCase__ ( cls): return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy())
76
a__ = '''0.18.2''' from .configuration_utils import ConfigMixin from .utils import ( OptionalDependencyNotAvailable, is_flax_available, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_note_seq_available, is_onnx_available, is_scipy_available, is_torch_available, is_torchsde_available, is_transformers_available, is_transformers_version, is_unidecode_available, logging, ) try: if not is_onnx_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_onnx_objects import * # noqa F403 else: from .pipelines import OnnxRuntimeModel try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_pt_objects import * # noqa F403 else: from .models import ( AutoencoderKL, ControlNetModel, ModelMixin, PriorTransformer, TaFilmDecoder, TransformeraDModel, UNetaDModel, UNetaDConditionModel, UNetaDModel, UNetaDConditionModel, VQModel, ) from .optimization import ( get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, get_scheduler, ) from .pipelines import ( AudioPipelineOutput, ConsistencyModelPipeline, DanceDiffusionPipeline, DDIMPipeline, DDPMPipeline, DiffusionPipeline, DiTPipeline, ImagePipelineOutput, KarrasVePipeline, LDMPipeline, LDMSuperResolutionPipeline, PNDMPipeline, RePaintPipeline, ScoreSdeVePipeline, ) from .schedulers import ( CMStochasticIterativeScheduler, DDIMInverseScheduler, DDIMParallelScheduler, DDIMScheduler, DDPMParallelScheduler, DDPMScheduler, DEISMultistepScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, HeunDiscreteScheduler, IPNDMScheduler, KarrasVeScheduler, KDPMaAncestralDiscreteScheduler, KDPMaDiscreteScheduler, PNDMScheduler, RePaintScheduler, SchedulerMixin, ScoreSdeVeScheduler, UnCLIPScheduler, UniPCMultistepScheduler, VQDiffusionScheduler, ) from .training_utils import EMAModel try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .schedulers import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .schedulers import DPMSolverSDEScheduler try: if not (is_torch_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipelines import ( AltDiffusionImgaImgPipeline, AltDiffusionPipeline, AudioLDMPipeline, CycleDiffusionPipeline, IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ImageTextPipelineOutput, KandinskyImgaImgPipeline, KandinskyInpaintPipeline, KandinskyPipeline, KandinskyPriorPipeline, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaControlnetPipeline, KandinskyVaaImgaImgPipeline, KandinskyVaaInpaintPipeline, KandinskyVaaPipeline, KandinskyVaaPriorEmbaEmbPipeline, KandinskyVaaPriorPipeline, LDMTextToImagePipeline, PaintByExamplePipeline, SemanticStableDiffusionPipeline, ShapEImgaImgPipeline, ShapEPipeline, StableDiffusionAttendAndExcitePipeline, StableDiffusionControlNetImgaImgPipeline, StableDiffusionControlNetInpaintPipeline, StableDiffusionControlNetPipeline, StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionImageVariationPipeline, StableDiffusionImgaImgPipeline, StableDiffusionInpaintPipeline, StableDiffusionInpaintPipelineLegacy, StableDiffusionInstructPixaPixPipeline, StableDiffusionLatentUpscalePipeline, StableDiffusionLDMaDPipeline, StableDiffusionModelEditingPipeline, StableDiffusionPanoramaPipeline, StableDiffusionParadigmsPipeline, StableDiffusionPipeline, StableDiffusionPipelineSafe, StableDiffusionPixaPixZeroPipeline, StableDiffusionSAGPipeline, StableDiffusionUpscalePipeline, StableUnCLIPImgaImgPipeline, StableUnCLIPPipeline, TextToVideoSDPipeline, TextToVideoZeroPipeline, UnCLIPImageVariationPipeline, UnCLIPPipeline, UniDiffuserModel, UniDiffuserPipeline, UniDiffuserTextDecoder, VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, VideoToVideoSDPipeline, VQDiffusionPipeline, ) try: if not (is_torch_available() and is_transformers_available() and is_invisible_watermark_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_invisible_watermark_objects import * # noqa F403 else: from .pipelines import StableDiffusionXLImgaImgPipeline, StableDiffusionXLPipeline try: if not (is_torch_available() and is_transformers_available() and is_k_diffusion_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipelines import StableDiffusionKDiffusionPipeline try: if not (is_torch_available() and is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_onnx_objects import * # noqa F403 else: from .pipelines import ( OnnxStableDiffusionImgaImgPipeline, OnnxStableDiffusionInpaintPipeline, OnnxStableDiffusionInpaintPipelineLegacy, OnnxStableDiffusionPipeline, OnnxStableDiffusionUpscalePipeline, StableDiffusionOnnxPipeline, ) try: if not (is_torch_available() and is_librosa_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_librosa_objects import * # noqa F403 else: from .pipelines import AudioDiffusionPipeline, Mel try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .pipelines import SpectrogramDiffusionPipeline try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_objects import * # noqa F403 else: from .models.controlnet_flax import FlaxControlNetModel from .models.modeling_flax_utils import FlaxModelMixin from .models.unet_ad_condition_flax import FlaxUNetaDConditionModel from .models.vae_flax import FlaxAutoencoderKL from .pipelines import FlaxDiffusionPipeline from .schedulers import ( FlaxDDIMScheduler, FlaxDDPMScheduler, FlaxDPMSolverMultistepScheduler, FlaxKarrasVeScheduler, FlaxLMSDiscreteScheduler, FlaxPNDMScheduler, FlaxSchedulerMixin, FlaxScoreSdeVeScheduler, ) try: if not (is_flax_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_and_transformers_objects import * # noqa F403 else: from .pipelines import ( FlaxStableDiffusionControlNetPipeline, FlaxStableDiffusionImgaImgPipeline, FlaxStableDiffusionInpaintPipeline, FlaxStableDiffusionPipeline, ) try: if not (is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_note_seq_objects import * # noqa F403 else: from .pipelines import MidiProcessor
235
0
def A__ ( __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = generate_pascal_triangle(__lowerCamelCase ) for row_idx in range(__lowerCamelCase ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=''' ''' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx], end=''' ''' ) else: print(triangle[row_idx][col_idx], end='''''' ) print() def A__ ( __lowerCamelCase ): if not isinstance(__lowerCamelCase, __lowerCamelCase ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) SCREAMING_SNAKE_CASE_ = [] for current_row_idx in range(__lowerCamelCase ): SCREAMING_SNAKE_CASE_ = populate_current_row(__lowerCamelCase, __lowerCamelCase ) triangle.append(__lowerCamelCase ) return triangle def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 1, 1 for current_col_idx in range(1, __lowerCamelCase ): calculate_current_element( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) return current_row def A__ ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, ): SCREAMING_SNAKE_CASE_ = triangle[current_row_idx - 1][current_col_idx - 1] SCREAMING_SNAKE_CASE_ = triangle[current_row_idx - 1][current_col_idx] SCREAMING_SNAKE_CASE_ = above_to_left_elt + above_to_right_elt def A__ ( __lowerCamelCase ): if not isinstance(__lowerCamelCase, __lowerCamelCase ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) SCREAMING_SNAKE_CASE_ = [[1]] for row_index in range(1, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [0] + result[-1] + [0] SCREAMING_SNAKE_CASE_ = row_index + 1 # Calculate the number of distinct elements in a row SCREAMING_SNAKE_CASE_ = sum(divmod(__lowerCamelCase, 2 ) ) SCREAMING_SNAKE_CASE_ = [ temp_row[i - 1] + temp_row[i] for i in range(1, distinct_elements + 1 ) ] SCREAMING_SNAKE_CASE_ = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() SCREAMING_SNAKE_CASE_ = row_first_half + row_second_half result.append(__lowerCamelCase ) return result def A__ ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(__lowerCamelCase, __lowerCamelCase ) -> None: SCREAMING_SNAKE_CASE_ = F'''{func.__name__}({value})''' SCREAMING_SNAKE_CASE_ = timeit(F'''__main__.{call}''', setup='''import __main__''' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(F'''{call:38} -- {timing:.4f} seconds''' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(__lowerCamelCase, __lowerCamelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
257
from graphs.minimum_spanning_tree_kruskal import kruskal def A__ ( ): SCREAMING_SNAKE_CASE_ = 9 SCREAMING_SNAKE_CASE_ = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] SCREAMING_SNAKE_CASE_ = kruskal(__lowerCamelCase, __lowerCamelCase ) SCREAMING_SNAKE_CASE_ = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] assert sorted(__lowerCamelCase ) == sorted(__lowerCamelCase )
257
1
"""simple docstring""" import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : List[str] = '''hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline''' def _lowercase ( self : Tuple , UpperCAmelCase__ : Union[str, Any]=0 ) -> int: _a : Dict = np.random.RandomState(UpperCAmelCase__ ) _a : str = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def _lowercase ( self : Any ) -> Union[str, Any]: _a : Any = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Any = self.get_dummy_inputs() _a : Union[str, Any] = pipe(**UpperCAmelCase__ ).images _a : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : Optional[int] = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : Union[str, Any] ) -> int: _a : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Dict = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[str] = self.get_dummy_inputs() _a : List[str] = pipe(**UpperCAmelCase__ ).images _a : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : str = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : int ) -> Any: _a : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Tuple = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Dict = self.get_dummy_inputs() _a : Any = pipe(**UpperCAmelCase__ ).images _a : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : Optional[int] ) -> str: _a : Tuple = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Union[str, Any] = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = self.get_dummy_inputs() _a : int = pipe(**UpperCAmelCase__ ).images _a : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : List[Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : Optional[int] ) -> int: _a : Tuple = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Union[str, Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[Any] = self.get_dummy_inputs() _a : Union[str, Any] = pipe(**UpperCAmelCase__ ).images _a : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : str = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : List[str] ) -> str: _a : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Union[str, Any] = self.get_dummy_inputs() _a : Any = pipe(**UpperCAmelCase__ ).images _a : int = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : Optional[int] = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : List[str] ) -> List[str]: _a : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[Any] = self.get_dummy_inputs() _a : Optional[int] = 3 * [inputs["""prompt"""]] # forward _a : Dict = pipe(**UpperCAmelCase__ ) _a : Tuple = output.images[0, -3:, -3:, -1] _a : str = self.get_dummy_inputs() _a : Union[str, Any] = 3 * [inputs.pop("""prompt""" )] _a : Any = pipe.tokenizer( UpperCAmelCase__ , padding="""max_length""" , max_length=pipe.tokenizer.model_max_length , truncation=UpperCAmelCase__ , return_tensors="""np""" , ) _a : Tuple = text_inputs["""input_ids"""] _a : Dict = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] _a : int = prompt_embeds # forward _a : List[str] = pipe(**UpperCAmelCase__ ) _a : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 def _lowercase ( self : Any ) -> int: _a : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Tuple = self.get_dummy_inputs() _a : List[str] = 3 * ["""this is a negative prompt"""] _a : Optional[Any] = negative_prompt _a : Optional[int] = 3 * [inputs["""prompt"""]] # forward _a : Any = pipe(**UpperCAmelCase__ ) _a : List[Any] = output.images[0, -3:, -3:, -1] _a : int = self.get_dummy_inputs() _a : int = 3 * [inputs.pop("""prompt""" )] _a : Tuple = [] for p in [prompt, negative_prompt]: _a : Optional[int] = pipe.tokenizer( UpperCAmelCase__ , padding="""max_length""" , max_length=pipe.tokenizer.model_max_length , truncation=UpperCAmelCase__ , return_tensors="""np""" , ) _a : Optional[int] = text_inputs["""input_ids"""] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] ) _a , _a : str = embeds # forward _a : List[str] = pipe(**UpperCAmelCase__ ) _a : Optional[int] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): @property def _lowercase ( self : int ) -> Optional[int]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _lowercase ( self : Optional[Any] ) -> List[Any]: _a : Optional[Any] = ort.SessionOptions() _a : Dict = False return options def _lowercase ( self : str ) -> Any: # using the PNDM scheduler by default _a : int = OnnxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = """A painting of a squirrel eating a burger""" np.random.seed(0 ) _a : Dict = sd_pipe([prompt] , guidance_scale=6.0 , num_inference_steps=10 , output_type="""np""" ) _a : Union[str, Any] = output.images _a : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[Any] = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowercase ( self : Union[str, Any] ) -> List[str]: _a : Any = DDIMScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) _a : str = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Tuple = """open neural network exchange""" _a : List[str] = np.random.RandomState(0 ) _a : Optional[int] = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase__ , output_type="""np""" ) _a : List[Any] = output.images _a : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _a : Dict = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowercase ( self : Optional[Any] ) -> List[Any]: _a : Union[str, Any] = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) _a : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Any = """open neural network exchange""" _a : List[str] = np.random.RandomState(0 ) _a : Optional[Any] = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase__ , output_type="""np""" ) _a : List[str] = output.images _a : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _a : List[str] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowercase ( self : Optional[int] ) -> str: _a : Optional[int] = 0 def test_callback_fn(UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : np.ndarray ) -> None: _a : Dict = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _a : str = latents[0, -3:, -3:, -1] _a : List[Any] = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _a : Optional[Any] = latents[0, -3:, -3:, -1] _a : List[Any] = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 _a : List[str] = False _a : Tuple = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Any = """Andromeda galaxy in a bottle""" _a : Dict = np.random.RandomState(0 ) pipe( prompt=UpperCAmelCase__ , num_inference_steps=5 , guidance_scale=7.5 , generator=UpperCAmelCase__ , callback=UpperCAmelCase__ , callback_steps=1 , ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def _lowercase ( self : List[str] ) -> Union[str, Any]: _a : List[str] = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) assert isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) assert pipe.safety_checker is None _a : List[Any] = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCAmelCase__ ) _a : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained(UpperCAmelCase__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None _a : List[Any] = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None
294
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _snake_case = logging.getLogger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=-1 ) -> Tuple: # in NER datasets, the last column is usually reserved for NER label _a : Optional[int] = label_idx def _lowercase ( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : Any = mode.value _a : Optional[int] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : int = 1 _a : int = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: _a : str = [] _a : str = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 _a : List[str] = [] _a : str = [] else: _a : List[Any] = line.split(""" """ ) words.append(splits[0] ) if len(UpperCAmelCase__ ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) return examples def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Union[str, Any]: _a : List[str] = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(UpperCAmelCase__ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _a : int = line.split()[0] + """ """ + preds_list[example_id].pop(0 ) + """\n""" writer.write(UpperCAmelCase__ ) else: logger.warning("""Maximum sequence length exceeded: No prediction for '%s'.""" , line.split()[0] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : List[Any] = f.read().splitlines() if "O" not in labels: _a : Union[str, Any] = ["""O"""] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class UpperCamelCase ( snake_case_ ): def __init__( self : Union[str, Any] ) -> List[str]: # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : Optional[int] = f.read().splitlines() if "O" not in labels: _a : Optional[Any] = ["""O"""] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : List[Any] = mode.value _a : Union[str, Any] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : List[str] = 1 _a : Optional[Any] = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: for sentence in parse_incr(UpperCAmelCase__ ): _a : List[Any] = [] _a : Any = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 return examples def _lowercase ( self : Tuple , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Dict: _a : Optional[Any] = 0 for sentence in parse_incr(UpperCAmelCase__ ): _a : List[str] = preds_list[example_id] _a : str = """""" for token in sentence: out += f"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(UpperCAmelCase__ ) example_id += 1 def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
294
1
'''simple docstring''' from __future__ import annotations import math __A : Union[str, Any] = "2020.9.26" __A : Union[str, Any] = "xcodz-dot, cclaus, dhruvmanila" def UpperCamelCase_ ( A__ : Optional[int] , A__ : List[Any] , A__ : Dict , A__ : str , A__ : int ): '''simple docstring''' if not all(isinstance(_lowerCAmelCase , (float, int) ) for val in locals().values() ): lowerCAmelCase_ : Tuple = f'Input values must either be float or int: {list(locals().values() )}' raise TypeError(_lowerCAmelCase ) lowerCAmelCase_ : List[str] = ((x * distance) / (z + distance)) * scale lowerCAmelCase_ : List[Any] = ((y * distance) / (z + distance)) * scale return projected_x, projected_y def UpperCamelCase_ ( A__ : str , A__ : Dict , A__ : int , A__ : Tuple , A__ : Optional[Any] ): '''simple docstring''' if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("""Axis must be a str""" ) lowerCAmelCase_ : List[str] = locals() del input_variables["axis"] if not all(isinstance(_lowerCAmelCase , (float, int) ) for val in input_variables.values() ): lowerCAmelCase_ : Union[str, Any] = ( """Input values except axis must either be float or int: """ f'{list(input_variables.values() )}' ) raise TypeError(_lowerCAmelCase ) lowerCAmelCase_ : Optional[int] = (angle % 3_60) / 4_50 * 1_80 / math.pi if axis == "z": lowerCAmelCase_ : Dict = x * math.cos(_lowerCAmelCase ) - y * math.sin(_lowerCAmelCase ) lowerCAmelCase_ : Any = y * math.cos(_lowerCAmelCase ) + x * math.sin(_lowerCAmelCase ) lowerCAmelCase_ : Union[str, Any] = z elif axis == "x": lowerCAmelCase_ : Tuple = y * math.cos(_lowerCAmelCase ) - z * math.sin(_lowerCAmelCase ) lowerCAmelCase_ : Any = z * math.cos(_lowerCAmelCase ) + y * math.sin(_lowerCAmelCase ) lowerCAmelCase_ : Optional[int] = x elif axis == "y": lowerCAmelCase_ : Tuple = x * math.cos(_lowerCAmelCase ) - z * math.sin(_lowerCAmelCase ) lowerCAmelCase_ : Optional[Any] = z * math.cos(_lowerCAmelCase ) + x * math.sin(_lowerCAmelCase ) lowerCAmelCase_ : List[str] = y else: raise ValueError("""not a valid axis, choose one of 'x', 'y', 'z'""" ) return new_x, new_y, new_z if __name__ == "__main__": import doctest doctest.testmod() print(F'''{convert_to_ad(1.0, 2.0, 3.0, 1_0.0, 1_0.0) = }''') print(F'''{rotate(1.0, 2.0, 3.0, 'y', 9_0.0) = }''')
371
'''simple docstring''' from __future__ import annotations def UpperCamelCase_ ( A__ : int | float | str , A__ : int | float | str ): '''simple docstring''' if nth_term == "": return [""] lowerCAmelCase_ : str = int(A__ ) lowerCAmelCase_ : Tuple = int(A__ ) lowerCAmelCase_ : list[str] = [] for temp in range(int(A__ ) ): series.append(f'1 / {pow(temp + 1 , int(A__ ) )}' if series else """1""" ) return series if __name__ == "__main__": import doctest doctest.testmod() __A : str = int(input("Enter the last number (nth term) of the P-Series")) __A : Tuple = int(input("Enter the power for P-Series")) print("Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p") print(p_series(nth_term, power))
89
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
319
from decimal import Decimal, getcontext from math import ceil, factorial def UpperCAmelCase ( lowercase ): """simple docstring""" if not isinstance(lowercase , lowercase ): raise TypeError('''Undefined for non-integers''' ) elif precision < 1: raise ValueError('''Undefined for non-natural numbers''' ) __lowercase = precision __lowercase = ceil(precision / 14 ) __lowercase = 426880 * Decimal(10005 ).sqrt() __lowercase = 1 __lowercase = 13591409 __lowercase = Decimal(lowercase ) for k in range(1 , lowercase ): __lowercase = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowercase ) ** 3) linear_term += 545140134 exponential_term *= -262537412640768000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __a : Optional[Any] = 5_0 print(F'''The first {n} digits of pi is: {pi(n)}''')
210
0
'''simple docstring''' import math from datetime import datetime, timedelta def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : List[Any] = year % 19 lowerCAmelCase__ : Dict = year % 4 lowerCAmelCase__ : Dict = year % 7 lowerCAmelCase__ : int = math.floor(year / 100 ) lowerCAmelCase__ : List[str] = math.floor((13 + 8 * leap_day_inhibits) / 25 ) lowerCAmelCase__ : Dict = leap_day_inhibits / 4 lowerCAmelCase__ : Optional[Any] = ( 15 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 30 lowerCAmelCase__ : Union[str, Any] = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 lowerCAmelCase__ : int = (19 * metonic_cycle + secular_moon_shift) % 30 # PHM -> Paschal Full Moon lowerCAmelCase__ : List[Any] = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 29 and days_from_phm_to_sunday == 6: return datetime(UpperCamelCase , 4 , 19 ) elif days_to_add == 28 and days_from_phm_to_sunday == 6: return datetime(UpperCamelCase , 4 , 18 ) else: return datetime(UpperCamelCase , 3 , 22 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (1994, 2000, 2010, 2021, 2023): _lowerCAmelCase = '''will be''' if year > datetime.now().year else '''was''' print(F"""Easter in {year} {tense} {gauss_easter(year)}""")
360
'''simple docstring''' import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: _lowerCAmelCase = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class lowerCAmelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase=7 ,__UpperCAmelCase=3 ,__UpperCAmelCase=18 ,__UpperCAmelCase=30 ,__UpperCAmelCase=400 ,__UpperCAmelCase=None ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=None ,) -> Optional[Any]: lowerCAmelCase__ : Optional[Any] = size if size is not None else {"""height""": 20, """width""": 20} lowerCAmelCase__ : List[Any] = parent lowerCAmelCase__ : List[str] = batch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = image_size lowerCAmelCase__ : Optional[Any] = min_resolution lowerCAmelCase__ : Tuple = max_resolution lowerCAmelCase__ : List[Any] = size lowerCAmelCase__ : List[str] = do_normalize lowerCAmelCase__ : Optional[int] = do_convert_rgb lowerCAmelCase__ : str = [512, 1024, 2048, 4096] lowerCAmelCase__ : int = patch_size if patch_size is not None else {"""height""": 16, """width""": 16} def UpperCAmelCase_ ( self ) -> Optional[int]: return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : Optional[Any] = """https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg""" lowerCAmelCase__ : Union[str, Any] = Image.open(requests.get(__UpperCAmelCase ,stream=__UpperCAmelCase ).raw ).convert("""RGB""" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): '''simple docstring''' __lowercase : Union[str, Any] = PixaStructImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> List[str]: lowerCAmelCase__ : Union[str, Any] = PixaStructImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[Any]: lowerCAmelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_normalize""" ) ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_convert_rgb""" ) ) def UpperCAmelCase_ ( self ) -> Optional[int]: lowerCAmelCase__ : List[Any] = self.image_processor_tester.prepare_dummy_image() lowerCAmelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) lowerCAmelCase__ : str = 2048 lowerCAmelCase__ : Tuple = image_processor(__UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() ,torch.tensor(0.0_6_0_6 ) ,atol=1E-3 ,rtol=1E-3 ) ) def UpperCAmelCase_ ( self ) -> str: # Initialize image_processor lowerCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,Image.Image ) # Test not batched input lowerCAmelCase__ : List[str] = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : str = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : Any = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) def UpperCAmelCase_ ( self ) -> Any: # Initialize image_processor lowerCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,Image.Image ) # Test not batched input lowerCAmelCase__ : Tuple = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 lowerCAmelCase__ : Optional[int] = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(__UpperCAmelCase ): lowerCAmelCase__ : Any = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches lowerCAmelCase__ : Optional[Any] = """Hello""" lowerCAmelCase__ : List[str] = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ,header_text=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : str = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ,header_text=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image_processor lowerCAmelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase__ : List[str] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ,numpify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,np.ndarray ) lowerCAmelCase__ : Tuple = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : Any = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : int = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) def UpperCAmelCase_ ( self ) -> Optional[int]: # Initialize image_processor lowerCAmelCase__ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ,torchify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,torch.Tensor ) # Test not batched input lowerCAmelCase__ : Optional[int] = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : Dict = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : Union[str, Any] = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): '''simple docstring''' __lowercase : Optional[int] = PixaStructImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> List[Any]: lowerCAmelCase__ : Dict = PixaStructImageProcessingTester(self ,num_channels=4 ) lowerCAmelCase__ : str = 3 @property def UpperCAmelCase_ ( self ) -> Tuple: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_normalize""" ) ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_convert_rgb""" ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: # Initialize image_processor lowerCAmelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,Image.Image ) # Test not batched input lowerCAmelCase__ : Dict = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : int = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : Dict = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,)
184
0
import unittest import numpy as np def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , ) -> np.ndarray: """simple docstring""" A : Tuple = np.shape(lowercase__ ) A : Any = np.shape(lowercase__ ) A : Dict = np.shape(lowercase__ ) if shape_a[0] != shape_b[0]: A : Any = ( """Expected the same number of rows for A and B. """ f'''Instead found A of size {shape_a} and B of size {shape_b}''' ) raise ValueError(lowercase__ ) if shape_b[1] != shape_c[1]: A : Optional[int] = ( """Expected the same number of columns for B and C. """ f'''Instead found B of size {shape_b} and C of size {shape_c}''' ) raise ValueError(lowercase__ ) A : Tuple = pseudo_inv if a_inv is None: try: A : Optional[Any] = np.linalg.inv(lowercase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ): A : int = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) A : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) A : List[str] = np.array([[2, 1], [6, 3]] ) A : List[Any] = schur_complement(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) A : Dict = np.block([[a, b], [b.T, c]] ) A : Union[str, Any] = np.linalg.det(lowerCamelCase__ ) A : Dict = np.linalg.det(lowerCamelCase__ ) A : str = np.linalg.det(lowerCamelCase__ ) self.assertAlmostEqual(lowerCamelCase__, det_a * det_s ) def _lowerCAmelCase ( self ): A : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) A : Optional[Any] = np.array([[0, 3], [3, 0], [2, 3]] ) A : Any = np.array([[2, 1], [6, 3]] ) with self.assertRaises(lowerCamelCase__ ): schur_complement(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) def _lowerCAmelCase ( self ): A : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) A : List[Any] = np.array([[0, 3], [3, 0], [2, 3]] ) A : Dict = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(lowerCamelCase__ ): schur_complement(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
116
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase = '''▁''' lowerCAmelCase = {'''vocab_file''': '''spiece.model'''} lowerCAmelCase = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } lowerCAmelCase = { '''google/pegasus-xsum''': 5_1_2, } lowerCAmelCase = logging.get_logger(__name__) class A ( A_ ): UpperCamelCase_ : Union[str, Any] =VOCAB_FILES_NAMES UpperCamelCase_ : List[Any] =VOCAB_FILES_NAMES UpperCamelCase_ : int =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ : Tuple =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ : int =['''input_ids''', '''attention_mask'''] def __init__(self , lowerCAmelCase , lowerCAmelCase="<pad>" , lowerCAmelCase="</s>" , lowerCAmelCase="<unk>" , lowerCAmelCase="<mask_2>" , lowerCAmelCase="<mask_1>" , lowerCAmelCase=None , lowerCAmelCase=1_0_3 , lowerCAmelCase = None , **lowerCAmelCase , ): __lowercase= offset if additional_special_tokens is not None: if not isinstance(lowerCAmelCase , lowerCAmelCase ): raise TypeError( f'additional_special_tokens should be of type {type(lowerCAmelCase )}, but is' f' {type(lowerCAmelCase )}' ) __lowercase= ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(lowerCAmelCase ) , self.offset - 1 ) ] if len(set(lowerCAmelCase ) ) != len(lowerCAmelCase ): raise ValueError( 'Please make sure that the provided additional_special_tokens do not contain an incorrectly' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) __lowercase= additional_special_tokens_extended else: __lowercase= [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] __lowercase= {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowerCAmelCase , unk_token=lowerCAmelCase , mask_token=lowerCAmelCase , pad_token=lowerCAmelCase , mask_token_sent=lowerCAmelCase , offset=lowerCAmelCase , additional_special_tokens=lowerCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase , ) __lowercase= mask_token_sent __lowercase= vocab_file __lowercase= spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCAmelCase ) # add special tokens to encoder dict __lowercase= { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) __lowercase= {v: k for k, v in self.encoder.items()} @property def _A (self ): return len(self.sp_model ) + self.offset def _A (self ): __lowercase= {self.convert_ids_to_tokens(lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__(self ): __lowercase= self.__dict__.copy() __lowercase= None return state def __setstate__(self , lowerCAmelCase ): __lowercase= d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __lowercase= {} __lowercase= spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A (self , lowerCAmelCase ): return self.sp_model.encode(lowerCAmelCase , out_type=lowerCAmelCase ) def _A (self , lowerCAmelCase ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] __lowercase= self.sp_model.piece_to_id(lowerCAmelCase ) return sp_id + self.offset def _A (self , lowerCAmelCase ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: __lowercase= self.sp_model.IdToPiece(index - self.offset ) return token def _A (self , lowerCAmelCase ): __lowercase= [] __lowercase= '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCAmelCase ) + token __lowercase= [] else: current_sub_tokens.append(lowerCAmelCase ) out_string += self.sp_model.decode(lowerCAmelCase ) return out_string.strip() def _A (self , lowerCAmelCase=False ): return 1 def _A (self , lowerCAmelCase ): __lowercase= set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A (self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = False ): if already_has_special_tokens: return self._special_token_mask(lowerCAmelCase ) elif token_ids_a is None: return self._special_token_mask(lowerCAmelCase ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A (self , lowerCAmelCase , lowerCAmelCase=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A (self , lowerCAmelCase , lowerCAmelCase = None ): if not os.path.isdir(lowerCAmelCase ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase= os.path.join( lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase , 'wb' ) as fi: __lowercase= self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase ) return (out_vocab_file,)
295
0
from __future__ import annotations def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): lowercase :list[list[int]] = [] lowercase :list[int] = [] lowercase :Optional[Any] = 0 lowercase :Any = sum(lowerCamelCase ) create_state_space_tree(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) return result def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ): if sum(lowerCamelCase ) > max_sum or (remaining_nums_sum + sum(lowerCamelCase )) < max_sum: return if sum(lowerCamelCase ) == max_sum: result.append(lowerCamelCase ) return for index in range(lowerCamelCase, len(lowerCamelCase ) ): create_state_space_tree( lowerCamelCase, lowerCamelCase, index + 1, [*path, nums[index]], lowerCamelCase, remaining_nums_sum - nums[index], ) _UpperCAmelCase : int = [3, 34, 4, 12, 5, 2] _UpperCAmelCase : Tuple = 9 _UpperCAmelCase : List[str] = generate_sum_of_subsets_soln(nums, max_sum) print(*result)
158
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, **lowerCamelCase ): lowercase :List[Any] = AutoConfig.from_pretrained(lowerCamelCase, **lowerCamelCase ) lowercase :Union[str, Any] = AutoModelForSeqaSeqLM.from_config(lowerCamelCase ) model.save_pretrained(lowerCamelCase ) AutoTokenizer.from_pretrained(lowerCamelCase ).save_pretrained(lowerCamelCase ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
158
1
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass snake_case : int = (3, 9, -11, 0, 7, 5, 1, -1) snake_case : Optional[Any] = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class _snake_case : SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 class _snake_case : def __init__( self , _lowerCamelCase ): a :Node | None = None for i in sorted(_lowerCamelCase , reverse=_lowerCamelCase ): a :int = Node(_lowerCamelCase , self.head ) def __iter__( self ): a :Dict = self.head while node: yield node.data a :List[str] = node.next_node def __len__( self ): return sum(1 for _ in self ) def __str__( self ): return " -> ".join([str(_lowerCamelCase ) for node in self] ) def __lowerCamelCase ( UpperCAmelCase_ : SortedLinkedList , UpperCAmelCase_ : SortedLinkedList ): """simple docstring""" return SortedLinkedList(list(UpperCAmelCase_ ) + list(UpperCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() snake_case : List[str] = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
94
def __lowerCamelCase ( UpperCAmelCase_ : int = 100_0000 ): """simple docstring""" a :Any = set(range(3 , UpperCAmelCase_ , 2 ) ) primes.add(2 ) for p in range(3 , UpperCAmelCase_ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , UpperCAmelCase_ , UpperCAmelCase_ ) ) ) a :Union[str, Any] = [float(UpperCAmelCase_ ) for n in range(limit + 1 )] for p in primes: for n in range(UpperCAmelCase_ , limit + 1 , UpperCAmelCase_ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F"""{solution() = }""")
94
1
from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class lowercase_ : _lowerCamelCase = field( metadata={'help': 'The output directory where the model will be written.'} , ) _lowerCamelCase = field( metadata={ 'help': ( 'The encoder model checkpoint for weights initialization.' 'Don\'t set if you want to train an encoder model from scratch.' ) } , ) _lowerCamelCase = field( metadata={ 'help': ( 'The decoder model checkpoint for weights initialization.' 'Don\'t set if you want to train a decoder model from scratch.' ) } , ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Pretrained encoder config name or path if not the same as encoder_model_name'} ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Pretrained decoder config name or path if not the same as decoder_model_name'} ) def snake_case () -> Optional[Any]: '''simple docstring''' _snake_case : List[str] = HfArgumentParser((ModelArguments,) ) (_snake_case) : List[Any] = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _snake_case : List[str] = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _snake_case : int = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _snake_case : Any = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _snake_case : Optional[Any] = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _snake_case : Tuple = True _snake_case : Optional[Any] = True _snake_case : Tuple = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=__lowercase , decoder_config=__lowercase , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _snake_case : Optional[int] = decoder_config.decoder_start_token_id _snake_case : List[str] = decoder_config.pad_token_id if decoder_start_token_id is None: _snake_case : List[str] = decoder_config.bos_token_id if pad_token_id is None: _snake_case : Tuple = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _snake_case : Optional[Any] = decoder_config.eos_token_id _snake_case : Optional[int] = decoder_start_token_id _snake_case : Dict = pad_token_id _snake_case : List[Any] = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _snake_case : Dict = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _snake_case : Union[str, Any] = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
365
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) class lowercase_ ( __snake_case ): _lowerCamelCase = ['pixel_values'] def __init__( self , lowercase_ = True , lowercase_ = None , lowercase_ = PILImageResampling.BILINEAR , lowercase_ = True , lowercase_ = 1 / 255 , lowercase_ = True , lowercase_ = None , lowercase_ = True , **lowercase_ , ): super().__init__(**lowercase_ ) _snake_case : Dict = size if size is not None else {"shortest_edge": 224} _snake_case : Optional[int] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) _snake_case : Any = crop_size if crop_size is not None else {"height": 256, "width": 256} _snake_case : str = get_size_dict(lowercase_ , param_name="crop_size" ) _snake_case : List[Any] = do_resize _snake_case : Tuple = size _snake_case : Union[str, Any] = resample _snake_case : str = do_rescale _snake_case : Dict = rescale_factor _snake_case : int = do_center_crop _snake_case : int = crop_size _snake_case : List[Any] = do_flip_channel_order def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ = PIL.Image.BILINEAR , lowercase_ = None , **lowercase_ , ): _snake_case : Optional[Any] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) if "shortest_edge" not in size: raise ValueError(f"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) _snake_case : int = get_resize_output_image_size(lowercase_ , size=size["shortest_edge"] , default_to_square=lowercase_ ) return resize(lowercase_ , size=lowercase_ , resample=lowercase_ , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ): _snake_case : List[str] = get_size_dict(lowercase_ ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(lowercase_ , size=(size["height"], size["width"]) , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ): return rescale(lowercase_ , scale=lowercase_ , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ = None ): return flip_channel_order(lowercase_ , data_format=lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = ChannelDimension.FIRST , **lowercase_ , ): _snake_case : Dict = do_resize if do_resize is not None else self.do_resize _snake_case : Union[str, Any] = resample if resample is not None else self.resample _snake_case : Dict = do_rescale if do_rescale is not None else self.do_rescale _snake_case : str = rescale_factor if rescale_factor is not None else self.rescale_factor _snake_case : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop _snake_case : List[str] = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) _snake_case : Optional[int] = size if size is not None else self.size _snake_case : Optional[int] = get_size_dict(lowercase_ , default_to_square=lowercase_ ) _snake_case : Optional[int] = crop_size if crop_size is not None else self.crop_size _snake_case : Union[str, Any] = get_size_dict(lowercase_ , param_name="crop_size" ) _snake_case : Tuple = make_list_of_images(lowercase_ ) if not valid_images(lowercase_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) # All transformations expect numpy arrays. _snake_case : Tuple = [to_numpy_array(lowercase_ ) for image in images] if do_resize: _snake_case : int = [self.resize(image=lowercase_ , size=lowercase_ , resample=lowercase_ ) for image in images] if do_center_crop: _snake_case : List[Any] = [self.center_crop(image=lowercase_ , size=lowercase_ ) for image in images] if do_rescale: _snake_case : Optional[int] = [self.rescale(image=lowercase_ , scale=lowercase_ ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: _snake_case : int = [self.flip_channel_order(image=lowercase_ ) for image in images] _snake_case : str = [to_channel_dimension_format(lowercase_ , lowercase_ ) for image in images] _snake_case : List[Any] = {"pixel_values": images} return BatchFeature(data=lowercase_ , tensor_type=lowercase_ ) def UpperCamelCase ( self , lowercase_ , lowercase_ = None ): _snake_case : int = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(lowercase_ ) != len(lowercase_ ): raise ValueError( "Make sure that you pass in as many target sizes as the batch dimension of the logits" ) if is_torch_tensor(lowercase_ ): _snake_case : Union[str, Any] = target_sizes.numpy() _snake_case : int = [] for idx in range(len(lowercase_ ) ): _snake_case : Optional[Any] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="bilinear" , align_corners=lowercase_ ) _snake_case : List[str] = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(lowercase_ ) else: _snake_case : List[Any] = logits.argmax(dim=1 ) _snake_case : Optional[Any] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
284
0
"""simple docstring""" import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class a : """simple docstring""" def __init__( self: int , UpperCamelCase: Optional[int] , UpperCamelCase: Any=13 , UpperCamelCase: List[str]=7 , UpperCamelCase: List[Any]=True , UpperCamelCase: str=True , UpperCamelCase: Dict=True , UpperCamelCase: List[str]=True , UpperCamelCase: List[str]=99 , UpperCamelCase: Dict=32 , UpperCamelCase: List[Any]=5 , UpperCamelCase: List[str]=4 , UpperCamelCase: Dict=37 , UpperCamelCase: List[Any]="gelu" , UpperCamelCase: Dict=0.1 , UpperCamelCase: Any=0.1 , UpperCamelCase: int=5_12 , UpperCamelCase: Tuple=16 , UpperCamelCase: str=2 , UpperCamelCase: Optional[Any]=0.02 , UpperCamelCase: Any=3 , UpperCamelCase: Any=4 , UpperCamelCase: Dict=None , ): """simple docstring""" A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_token_type_ids A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = type_sequence_label_size A__ = initializer_range A__ = num_labels A__ = num_choices A__ = scope def UpperCamelCase ( self: Tuple ): """simple docstring""" A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) A__ = None if self.use_token_type_ids: A__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ = None A__ = None A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ = ids_tensor([self.batch_size] , self.num_choices ) A__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self: int ): """simple docstring""" return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase_ , initializer_range=self.initializer_range , ) def UpperCamelCase ( self: Optional[Any] , UpperCamelCase: Any , UpperCamelCase: List[str] , UpperCamelCase: Union[str, Any] , UpperCamelCase: Union[str, Any] , UpperCamelCase: Tuple , UpperCamelCase: Tuple , UpperCamelCase: Optional[int] ): """simple docstring""" A__ = NystromformerModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ ) A__ = model(lowercase_ , token_type_ids=lowercase_ ) A__ = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self: int , UpperCamelCase: Tuple , UpperCamelCase: List[Any] , UpperCamelCase: Union[str, Any] , UpperCamelCase: List[str] , UpperCamelCase: List[Any] , UpperCamelCase: Optional[int] , UpperCamelCase: Any ): """simple docstring""" A__ = NystromformerForMaskedLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self: Tuple , UpperCamelCase: int , UpperCamelCase: List[Any] , UpperCamelCase: Any , UpperCamelCase: str , UpperCamelCase: Dict , UpperCamelCase: Any , UpperCamelCase: Tuple ): """simple docstring""" A__ = NystromformerForQuestionAnswering(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , start_positions=lowercase_ , end_positions=lowercase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self: Tuple , UpperCamelCase: List[str] , UpperCamelCase: List[Any] , UpperCamelCase: Union[str, Any] , UpperCamelCase: str , UpperCamelCase: Optional[Any] , UpperCamelCase: str , UpperCamelCase: int ): """simple docstring""" A__ = self.num_labels A__ = NystromformerForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self: List[Any] , UpperCamelCase: List[str] , UpperCamelCase: int , UpperCamelCase: List[str] , UpperCamelCase: Dict , UpperCamelCase: Tuple , UpperCamelCase: Union[str, Any] , UpperCamelCase: List[str] ): """simple docstring""" A__ = self.num_labels A__ = NystromformerForTokenClassification(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self: Optional[Any] , UpperCamelCase: Tuple , UpperCamelCase: Dict , UpperCamelCase: int , UpperCamelCase: List[str] , UpperCamelCase: Optional[Any] , UpperCamelCase: str , UpperCamelCase: Union[str, Any] ): """simple docstring""" A__ = self.num_choices A__ = NystromformerForMultipleChoice(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ = model( lowercase_ , attention_mask=lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase ( self: Dict ): """simple docstring""" A__ = self.prepare_config_and_inputs() ( A__ ) = config_and_inputs A__ = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class a ( _UpperCAmelCase, _UpperCAmelCase, unittest.TestCase ): """simple docstring""" UpperCAmelCase = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) UpperCAmelCase = ( { "feature-extraction": NystromformerModel, "fill-mask": NystromformerForMaskedLM, "question-answering": NystromformerForQuestionAnswering, "text-classification": NystromformerForSequenceClassification, "token-classification": NystromformerForTokenClassification, "zero-shot": NystromformerForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase = False UpperCAmelCase = False def UpperCamelCase ( self: int ): """simple docstring""" A__ = NystromformerModelTester(self ) A__ = ConfigTester(self , config_class=lowercase_ , hidden_size=37 ) def UpperCamelCase ( self: Tuple ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self: Dict ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def UpperCamelCase ( self: List[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ = type self.model_tester.create_and_check_model(*lowercase_ ) def UpperCamelCase ( self: Optional[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase_ ) def UpperCamelCase ( self: Optional[int] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase_ ) def UpperCamelCase ( self: Optional[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase_ ) def UpperCamelCase ( self: List[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase_ ) def UpperCamelCase ( self: List[Any] ): """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase_ ) @slow def UpperCamelCase ( self: Optional[int] ): """simple docstring""" for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = NystromformerModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) @require_torch class a ( unittest.TestCase ): """simple docstring""" @slow def UpperCamelCase ( self: Union[str, Any] ): """simple docstring""" A__ = NystromformerModel.from_pretrained("""uw-madison/nystromformer-512""" ) A__ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): A__ = model(lowercase_ )[0] A__ = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , lowercase_ ) A__ = torch.tensor( [[[-0.4_532, -0.0_936, 0.5_137], [-0.2_676, 0.0_628, 0.6_186], [-0.3_629, -0.1_726, 0.4_716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase_ , atol=1e-4 ) ) @slow def UpperCamelCase ( self: Optional[int] ): """simple docstring""" A__ = """the [MASK] of Belgium is Brussels""" A__ = AutoTokenizer.from_pretrained("""uw-madison/nystromformer-512""" ) A__ = NystromformerForMaskedLM.from_pretrained("""uw-madison/nystromformer-512""" ) A__ = tokenizer(lowercase_ , return_tensors="""pt""" ) with torch.no_grad(): A__ = model(encoding.input_ids ).logits A__ = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(lowercase_ ) , """capital""" )
335
'''simple docstring''' _lowercase : str = tuple[float, float, float] _lowercase : List[Any] = tuple[float, float, float] def lowerCamelCase ( UpperCAmelCase__ : Pointad , UpperCAmelCase__ : Pointad ) -> Vectorad: lowercase_ : List[str] = end_pointa[0] - end_pointa[0] lowercase_ : Union[str, Any] = end_pointa[1] - end_pointa[1] lowercase_ : List[Any] = end_pointa[2] - end_pointa[2] return (x, y, z) def lowerCamelCase ( UpperCAmelCase__ : Vectorad , UpperCAmelCase__ : Vectorad ) -> Vectorad: lowercase_ : List[Any] = ab[1] * ac[2] - ab[2] * ac[1] # *i lowercase_ : Union[str, Any] = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j lowercase_ : List[str] = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def lowerCamelCase ( UpperCAmelCase__ : Vectorad , UpperCAmelCase__ : int ) -> bool: return tuple(round(UpperCAmelCase__ , UpperCAmelCase__ ) for x in vector ) == (0, 0, 0) def lowerCamelCase ( UpperCAmelCase__ : Pointad , UpperCAmelCase__ : Pointad , UpperCAmelCase__ : Pointad , UpperCAmelCase__ : int = 10 ) -> bool: lowercase_ : Dict = create_vector(UpperCAmelCase__ , UpperCAmelCase__ ) lowercase_ : Optional[int] = create_vector(UpperCAmelCase__ , UpperCAmelCase__ ) return is_zero_vector(get_ad_vectors_cross(UpperCAmelCase__ , UpperCAmelCase__ ) , UpperCAmelCase__ )
239
0
"""simple docstring""" from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class lowerCAmelCase__( __lowercase ): '''simple docstring''' __snake_case = ['vqvae'] def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , ) -> List[Any]: super().__init__() self.register_modules(unet=__lowerCamelCase , scheduler=__lowerCamelCase , mel=__lowerCamelCase , vqvae=__lowerCamelCase ) def UpperCamelCase_ ( self ) -> int: return 5_0 if isinstance(self.scheduler , __lowerCamelCase ) else 1_0_0_0 @torch.no_grad() def __call__( self , __lowerCamelCase = 1 , __lowerCamelCase = None , __lowerCamelCase = None , __lowerCamelCase = 0 , __lowerCamelCase = 0 , __lowerCamelCase = None , __lowerCamelCase = None , __lowerCamelCase = 0 , __lowerCamelCase = 0 , __lowerCamelCase = None , __lowerCamelCase = 0 , __lowerCamelCase = None , __lowerCamelCase = None , __lowerCamelCase=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: _SCREAMING_SNAKE_CASE : List[str] = steps or self.get_default_steps() self.scheduler.set_timesteps(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: _SCREAMING_SNAKE_CASE : Optional[int] = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: _SCREAMING_SNAKE_CASE : Union[str, Any] = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=__lowerCamelCase , device=self.device , ) _SCREAMING_SNAKE_CASE : Union[str, Any] = noise _SCREAMING_SNAKE_CASE : Optional[int] = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(__lowerCamelCase , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = self.mel.audio_slice_to_image(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = np.frombuffer(input_image.tobytes() , dtype="uint8" ).reshape( (input_image.height, input_image.width) ) _SCREAMING_SNAKE_CASE : Optional[int] = (input_image / 2_5_5) * 2 - 1 _SCREAMING_SNAKE_CASE : List[Any] = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: _SCREAMING_SNAKE_CASE : Union[str, Any] = self.vqvae.encode(torch.unsqueeze(__lowerCamelCase , 0 ) ).latent_dist.sample( generator=__lowerCamelCase )[0] _SCREAMING_SNAKE_CASE : int = self.vqvae.config.scaling_factor * input_images if start_step > 0: _SCREAMING_SNAKE_CASE : List[Any] = self.scheduler.add_noise(__lowerCamelCase , __lowerCamelCase , self.scheduler.timesteps[start_step - 1] ) _SCREAMING_SNAKE_CASE : int = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) _SCREAMING_SNAKE_CASE : Optional[Any] = int(mask_start_secs * pixels_per_second ) _SCREAMING_SNAKE_CASE : Optional[int] = int(mask_end_secs * pixels_per_second ) _SCREAMING_SNAKE_CASE : Optional[Any] = self.scheduler.add_noise(__lowerCamelCase , __lowerCamelCase , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , __lowerCamelCase ): _SCREAMING_SNAKE_CASE : List[str] = self.unet(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase )["sample"] else: _SCREAMING_SNAKE_CASE : str = self.unet(__lowerCamelCase , __lowerCamelCase )["sample"] if isinstance(self.scheduler , __lowerCamelCase ): _SCREAMING_SNAKE_CASE : Union[str, Any] = self.scheduler.step( model_output=__lowerCamelCase , timestep=__lowerCamelCase , sample=__lowerCamelCase , eta=__lowerCamelCase , generator=__lowerCamelCase , )["prev_sample"] else: _SCREAMING_SNAKE_CASE : List[Any] = self.scheduler.step( model_output=__lowerCamelCase , timestep=__lowerCamelCase , sample=__lowerCamelCase , generator=__lowerCamelCase , )["prev_sample"] if mask is not None: if mask_start > 0: _SCREAMING_SNAKE_CASE : str = mask[:, step, :, :mask_start] if mask_end > 0: _SCREAMING_SNAKE_CASE : Dict = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance _SCREAMING_SNAKE_CASE : Optional[Any] = 1 / self.vqvae.config.scaling_factor * images _SCREAMING_SNAKE_CASE : Dict = self.vqvae.decode(__lowerCamelCase )["sample"] _SCREAMING_SNAKE_CASE : Union[str, Any] = (images / 2 + 0.5).clamp(0 , 1 ) _SCREAMING_SNAKE_CASE : Union[str, Any] = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() _SCREAMING_SNAKE_CASE : List[str] = (images * 2_5_5).round().astype("uint8" ) _SCREAMING_SNAKE_CASE : Tuple = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(__lowerCamelCase , mode="RGB" ).convert("L" ) for _ in images) ) _SCREAMING_SNAKE_CASE : Tuple = [self.mel.image_to_audio(__lowerCamelCase ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(__lowerCamelCase )[:, np.newaxis, :] ) , **ImagePipelineOutput(__lowerCamelCase ) ) @torch.no_grad() def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase = 5_0 ) -> np.ndarray: assert isinstance(self.scheduler , __lowerCamelCase ) self.scheduler.set_timesteps(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[int] = np.array( [np.frombuffer(image.tobytes() , dtype="uint8" ).reshape((1, image.height, image.width) ) for image in images] ) _SCREAMING_SNAKE_CASE : Union[str, Any] = (sample / 2_5_5) * 2 - 1 _SCREAMING_SNAKE_CASE : Any = torch.Tensor(__lowerCamelCase ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): _SCREAMING_SNAKE_CASE : Optional[int] = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps _SCREAMING_SNAKE_CASE : Optional[Any] = self.scheduler.alphas_cumprod[t] _SCREAMING_SNAKE_CASE : List[str] = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) _SCREAMING_SNAKE_CASE : Optional[int] = 1 - alpha_prod_t _SCREAMING_SNAKE_CASE : Optional[int] = self.unet(__lowerCamelCase , __lowerCamelCase )["sample"] _SCREAMING_SNAKE_CASE : List[str] = (1 - alpha_prod_t_prev) ** 0.5 * model_output _SCREAMING_SNAKE_CASE : str = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) _SCREAMING_SNAKE_CASE : List[str] = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def UpperCamelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> torch.Tensor: _SCREAMING_SNAKE_CASE : Any = acos(torch.dot(torch.flatten(__lowerCamelCase ) , torch.flatten(__lowerCamelCase ) ) / torch.norm(__lowerCamelCase ) / torch.norm(__lowerCamelCase ) ) return sin((1 - alpha) * theta ) * xa / sin(__lowerCamelCase ) + sin(alpha * theta ) * xa / sin(__lowerCamelCase )
361
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ =np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ =[0, 25, 50] UpperCamelCase__ =[25, 50, 75] UpperCamelCase__ =fuzz.membership.trimf(X, abca) UpperCamelCase__ =fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ =np.ones(75) UpperCamelCase__ =np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ =fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ =fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ =fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ =fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ =young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ =young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ =fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ =fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
325
0
'''simple docstring''' from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class _lowerCAmelCase ( __UpperCAmelCase ): @slow @require_torch def _a (self ): A_ : List[str] = EncoderDecoderModel.from_encoder_decoder_pretrained("""prajjwal1/bert-tiny""" , """prajjwal1/bert-tiny""" ) A_ : List[str] = BertTokenizer.from_pretrained("""bert-base-uncased""" ) A_ : Any = bertabert.config.encoder.vocab_size A_ : Any = tokenizer.sep_token_id A_ : Tuple = tokenizer.cls_token_id A_ : str = 128 A_ : Optional[int] = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""train[:1%]""" ) A_ : Optional[Any] = datasets.load_dataset("""cnn_dailymail""" , """3.0.0""" , split="""validation[:1%]""" ) A_ : Tuple = train_dataset.select(range(32 ) ) A_ : List[str] = val_dataset.select(range(16 ) ) A_ : int = 4 def _map_to_encoder_decoder_inputs(lowercase ): # Tokenizer will automatically set [BOS] <text> [EOS] A_ : Tuple = tokenizer(batch["""article"""] , padding="""max_length""" , truncation=lowercase , max_length=512 ) A_ : str = tokenizer(batch["""highlights"""] , padding="""max_length""" , truncation=lowercase , max_length=128 ) A_ : List[str] = inputs.input_ids A_ : Tuple = inputs.attention_mask A_ : Tuple = outputs.input_ids A_ : List[Any] = outputs.input_ids.copy() A_ : List[str] = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["""labels"""] ] A_ : Optional[Any] = outputs.attention_mask assert all(len(lowercase ) == 512 for x in inputs.input_ids ) assert all(len(lowercase ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(lowercase ): A_ : Union[str, Any] = pred.label_ids A_ : str = pred.predictions # all unnecessary tokens are removed A_ : int = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase ) A_ : Optional[Any] = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase ) A_ : int = sum([int(pred_str[i] == label_str[i] ) for i in range(len(lowercase ) )] ) / len(lowercase ) return {"accuracy": accuracy} # map train dataset A_ : Tuple = train_dataset.map( _map_to_encoder_decoder_inputs , batched=lowercase , batch_size=lowercase , remove_columns=["""article""", """highlights"""] , ) train_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) # same for validation dataset A_ : int = val_dataset.map( _map_to_encoder_decoder_inputs , batched=lowercase , batch_size=lowercase , remove_columns=["""article""", """highlights"""] , ) val_dataset.set_format( type="""torch""" , columns=["""input_ids""", """attention_mask""", """decoder_input_ids""", """decoder_attention_mask""", """labels"""] , ) A_ : Tuple = self.get_auto_remove_tmp_dir() A_ : Dict = SeqaSeqTrainingArguments( output_dir=lowercase , per_device_train_batch_size=lowercase , per_device_eval_batch_size=lowercase , predict_with_generate=lowercase , evaluation_strategy="""steps""" , do_train=lowercase , do_eval=lowercase , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer A_ : Any = SeqaSeqTrainer( model=lowercase , args=lowercase , compute_metrics=_compute_metrics , train_dataset=lowercase , eval_dataset=lowercase , tokenizer=lowercase , ) # start training trainer.train()
206
'''simple docstring''' import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCAmelCase : def __init__(self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=3 , lowercase=4 , lowercase=None , ): A_ : str = parent A_ : int = batch_size A_ : Dict = seq_length A_ : Any = is_training A_ : List[str] = use_input_mask A_ : Any = use_token_type_ids A_ : int = use_labels A_ : str = vocab_size A_ : Any = hidden_size A_ : Optional[Any] = num_hidden_layers A_ : Dict = num_attention_heads A_ : Dict = intermediate_size A_ : Optional[int] = hidden_act A_ : int = hidden_dropout_prob A_ : Optional[int] = attention_probs_dropout_prob A_ : Any = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : Any = type_sequence_label_size A_ : Tuple = initializer_range A_ : int = num_labels A_ : Optional[int] = num_choices A_ : Optional[int] = scope def _a (self ): A_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : List[str] = None if self.use_input_mask: A_ : Any = random_attention_mask([self.batch_size, self.seq_length] ) A_ : int = None if self.use_token_type_ids: A_ : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : Tuple = None A_ : List[str] = None A_ : List[str] = None if self.use_labels: A_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : str = ids_tensor([self.batch_size] , self.num_choices ) A_ : Optional[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _a (self ): return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : List[Any] = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() A_ : List[Any] = model(lowercase , attention_mask=lowercase ) A_ : str = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ): A_ : List[Any] = BioGptForCausalLM(config=lowercase ) model.to(lowercase ) model.eval() A_ : Dict = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): A_ : Optional[Any] = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() # create attention mask A_ : str = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) A_ : Optional[Any] = self.seq_length // 2 A_ : List[Any] = 0 # first forward pass A_, A_ : List[str] = model(lowercase , attention_mask=lowercase ).to_tuple() # create hypothetical next token and extent to next_input_ids A_ : Optional[Any] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids A_ : Union[str, Any] = ids_tensor((1,) , lowercase ).item() + 1 A_ : List[str] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) A_ : Optional[int] = random_other_next_tokens # append to next input_ids and attn_mask A_ : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) A_ : Any = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=lowercase )] , dim=1 , ) # get two different outputs A_ : List[Any] = model(lowercase , attention_mask=lowercase )["""last_hidden_state"""] A_ : Optional[int] = model(lowercase , past_key_values=lowercase , attention_mask=lowercase )["""last_hidden_state"""] # select random slice A_ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A_ : int = output_from_no_past[:, -1, random_slice_idx].detach() A_ : List[Any] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-3 ) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): A_ : Optional[int] = BioGptModel(config=lowercase ).to(lowercase ).eval() A_ : Union[str, Any] = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) # first forward pass A_ : int = model(lowercase , attention_mask=lowercase , use_cache=lowercase ) A_, A_ : int = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids A_ : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A_ : str = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and A_ : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) A_ : int = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) A_ : List[str] = model(lowercase , attention_mask=lowercase )["""last_hidden_state"""] A_ : Tuple = model(lowercase , attention_mask=lowercase , past_key_values=lowercase )[ """last_hidden_state""" ] # select random slice A_ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A_ : List[str] = output_from_no_past[:, -3:, random_slice_idx].detach() A_ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-3 ) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase , lowercase=False ): A_ : Union[str, Any] = BioGptForCausalLM(lowercase ) model.to(lowercase ) if gradient_checkpointing: model.gradient_checkpointing_enable() A_ : Dict = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _a (self , lowercase , *lowercase ): A_ : Union[str, Any] = BioGptModel(lowercase ) A_ : str = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_01 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): A_ : Union[str, Any] = self.num_labels A_ : Optional[int] = BioGptForTokenClassification(lowercase ) model.to(lowercase ) model.eval() A_ : Dict = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a (self ): A_ : List[Any] = self.prepare_config_and_inputs() ( ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ) : Dict = config_and_inputs A_ : Tuple = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : str = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE : int = (BioGptForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE : Optional[Any] = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE : List[str] = False def _a (self ): A_ : Tuple = BioGptModelTester(self ) A_ : List[Any] = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def _a (self ): self.config_tester.run_common_tests() def _a (self ): A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def _a (self ): A_ : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A_ : int = type self.model_tester.create_and_check_model(*lowercase ) def _a (self ): A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*lowercase ) def _a (self ): A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*lowercase , gradient_checkpointing=lowercase ) def _a (self ): A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*lowercase ) def _a (self ): A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*lowercase ) def _a (self ): A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*lowercase ) @slow def _a (self ): A_ : str = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(lowercase ) A_ : List[str] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A_ : List[str] = """left""" # Define PAD Token = EOS Token = 50256 A_ : Any = tokenizer.eos_token A_ : Dict = model.config.eos_token_id # use different length sentences to test batching A_ : List[Any] = [ """Hello, my dog is a little""", """Today, I""", ] A_ : List[str] = tokenizer(lowercase , return_tensors="""pt""" , padding=lowercase ) A_ : List[str] = inputs["""input_ids"""].to(lowercase ) A_ : List[Any] = model.generate( input_ids=lowercase , attention_mask=inputs["""attention_mask"""].to(lowercase ) , ) A_ : List[str] = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(lowercase ) A_ : List[Any] = model.generate(input_ids=lowercase ) A_ : List[str] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() A_ : Union[str, Any] = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(lowercase ) A_ : Any = model.generate(input_ids=lowercase , max_length=model.config.max_length - num_paddings ) A_ : List[str] = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase ) A_ : str = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowercase ) A_ : List[str] = tokenizer.decode(output_padded[0] , skip_special_tokens=lowercase ) A_ : Union[str, Any] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(lowercase , lowercase ) self.assertListEqual(lowercase , [non_padded_sentence, padded_sentence] ) @slow def _a (self ): for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : int = BioGptModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def _a (self ): A_, A_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[Any] = 3 A_ : Union[str, Any] = input_dict["""input_ids"""] A_ : List[Any] = input_ids.ne(1 ).to(lowercase ) A_ : str = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A_ : Union[str, Any] = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() A_ : Tuple = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _a (self ): A_, A_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A_ : Tuple = 3 A_ : Dict = """multi_label_classification""" A_ : List[Any] = input_dict["""input_ids"""] A_ : Tuple = input_ids.ne(1 ).to(lowercase ) A_ : List[str] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A_ : Dict = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() A_ : int = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): @slow def _a (self ): A_ : Dict = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) A_ : Optional[Any] = torch.tensor([[2, 4805, 9, 656, 21]] ) A_ : Dict = model(lowercase )[0] A_ : Any = 42384 A_ : Union[str, Any] = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , lowercase ) A_ : Union[str, Any] = torch.tensor( [[[-9.52_36, -9.89_18, 10.45_57], [-11.04_69, -9.64_23, 8.10_22], [-8.86_64, -7.88_26, 5.53_25]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=1E-4 ) ) @slow def _a (self ): A_ : Union[str, Any] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A_ : Dict = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(lowercase ) torch.manual_seed(0 ) A_ : Dict = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(lowercase ) A_ : Optional[int] = model.generate( **lowercase , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=lowercase , ) A_ : Union[str, Any] = tokenizer.decode(output_ids[0] , skip_special_tokens=lowercase ) A_ : Any = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(lowercase , lowercase )
206
1
import os # Precomputes a list of the 100 first triangular numbers lowerCamelCase_ : Optional[int] = [int(0.5 * n * (n + 1)) for n in range(1, 1_01)] def A__ ( ) -> Dict: UpperCamelCase_: Optional[int] = os.path.dirname(os.path.realpath(__UpperCAmelCase ) ) UpperCamelCase_: Any = os.path.join(__UpperCAmelCase , """words.txt""" ) UpperCamelCase_: str = """""" with open(__UpperCAmelCase ) as f: UpperCamelCase_: str = f.readline() UpperCamelCase_: Optional[int] = [word.strip("""\"""" ) for word in words.strip("""\r\n""" ).split(""",""" )] UpperCamelCase_: int = [ word for word in [sum(ord(__UpperCAmelCase ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(__UpperCAmelCase ) if __name__ == "__main__": print(solution())
356
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline lowerCamelCase_ : Optional[Any] = datasets.utils.logging.get_logger(__name__) @dataclass class _UpperCamelCase ( datasets.BuilderConfig ): '''simple docstring''' __UpperCamelCase : Optional[datasets.Features] = None __UpperCamelCase : str = "utf-8" __UpperCamelCase : Optional[str] = None __UpperCamelCase : Optional[str] = None __UpperCamelCase : bool = True # deprecated __UpperCamelCase : Optional[int] = None # deprecated __UpperCamelCase : int = 10 << 20 # 10MB __UpperCamelCase : Optional[bool] = None class _UpperCamelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __UpperCamelCase : Tuple = JsonConfig def lowerCAmelCase__ ( self : int ): if self.config.block_size is not None: logger.warning("""The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead""" ) UpperCamelCase_: List[str] = self.config.block_size if self.config.use_threads is not True: logger.warning( """The JSON loader parameter `use_threads` is deprecated and doesn't have any effect anymore.""" ) if self.config.newlines_in_values is not None: raise ValueError("""The JSON loader parameter `newlines_in_values` is no longer supported""" ) return datasets.DatasetInfo(features=self.config.features ) def lowerCAmelCase__ ( self : Dict , snake_case_ : str ): if not self.config.data_files: raise ValueError(f'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) UpperCamelCase_: Dict = dl_manager.download_and_extract(self.config.data_files ) if isinstance(snake_case_ , (str, list, tuple) ): UpperCamelCase_: List[Any] = data_files if isinstance(snake_case_ , snake_case_ ): UpperCamelCase_: str = [files] UpperCamelCase_: Any = [dl_manager.iter_files(snake_case_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] UpperCamelCase_: Dict = [] for split_name, files in data_files.items(): if isinstance(snake_case_ , snake_case_ ): UpperCamelCase_: Tuple = [files] UpperCamelCase_: Optional[int] = [dl_manager.iter_files(snake_case_ ) for file in files] splits.append(datasets.SplitGenerator(name=snake_case_ , gen_kwargs={"""files""": files} ) ) return splits def lowerCAmelCase__ ( self : str , snake_case_ : pa.Table ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): UpperCamelCase_: Union[str, Any] = self.config.features.arrow_schema.field(snake_case_ ).type UpperCamelCase_: Tuple = pa_table.append_column(snake_case_ , pa.array([None] * len(snake_case_ ) , type=snake_case_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example UpperCamelCase_: int = table_cast(snake_case_ , self.config.features.arrow_schema ) return pa_table def lowerCAmelCase__ ( self : Dict , snake_case_ : Optional[Any] ): for file_idx, file in enumerate(itertools.chain.from_iterable(snake_case_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(snake_case_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: UpperCamelCase_: Dict = json.load(snake_case_ ) # We keep only the field we are interested in UpperCamelCase_: Optional[int] = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(snake_case_ , (list, tuple) ): UpperCamelCase_: Optional[int] = set().union(*[row.keys() for row in dataset] ) UpperCamelCase_: int = {col: [row.get(snake_case_ ) for row in dataset] for col in keys} else: UpperCamelCase_: Optional[int] = dataset UpperCamelCase_: List[str] = pa.Table.from_pydict(snake_case_ ) yield file_idx, self._cast_table(snake_case_ ) # If the file has one json object per line else: with open(snake_case_ , """rb""" ) as f: UpperCamelCase_: Optional[int] = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small UpperCamelCase_: Optional[int] = max(self.config.chunksize // 32 , 16 << 10 ) UpperCamelCase_: Tuple = ( self.config.encoding_errors if self.config.encoding_errors is not None else """strict""" ) while True: UpperCamelCase_: int = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(snake_case_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": UpperCamelCase_: Tuple = batch.decode(self.config.encoding , errors=snake_case_ ).encode("""utf-8""" ) try: while True: try: UpperCamelCase_: Tuple = paj.read_json( io.BytesIO(snake_case_ ) , read_options=paj.ReadOptions(block_size=snake_case_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(snake_case_ , pa.ArrowInvalid ) and "straddling" not in str(snake_case_ ) or block_size > len(snake_case_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( f'''Batch of {len(snake_case_ )} bytes couldn\'t be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.''' ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( snake_case_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: UpperCamelCase_: Optional[Any] = json.load(snake_case_ ) except json.JSONDecodeError: logger.error(f'''Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}''' ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(snake_case_ , snake_case_ ): # list is the only sequence type supported in JSON try: UpperCamelCase_: Any = set().union(*[row.keys() for row in dataset] ) UpperCamelCase_: List[str] = {col: [row.get(snake_case_ ) for row in dataset] for col in keys} UpperCamelCase_: int = pa.Table.from_pydict(snake_case_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(f'''Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}''' ) raise ValueError(f'''Not able to read records in the JSON file at {file}.''' ) from None yield file_idx, self._cast_table(snake_case_ ) break else: logger.error(f'''Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}''' ) raise ValueError( f'''Not able to read records in the JSON file at {file}. ''' f'''You should probably indicate the field of the JSON file containing your records. ''' f'''This JSON file contain the following fields: {str(list(dataset.keys() ) )}. ''' f'''Select the correct one and provide it as `field=\'XXX\'` to the dataset loading method. ''' ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(snake_case_ ) batch_idx += 1
223
0
"""simple docstring""" from math import sqrt def __lowerCAmelCase (_UpperCamelCase = 100_0000 ): __lowerCAmelCase : int = 0 __lowerCAmelCase : int = 0 __lowerCAmelCase : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'{solution() = }')
86
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class UpperCAmelCase ( A_ ): A__ : jnp.ndarray @flax_register_to_config class UpperCAmelCase ( nn.Module ,A_ ,A_ ): A__ : int = 32 A__ : int = 4 A__ : int = 4 A__ : Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) A__ : Tuple[str] = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") A__ : Union[bool, Tuple[bool]] = False A__ : Tuple[int] = (3_20, 6_40, 12_80, 12_80) A__ : int = 2 A__ : Union[int, Tuple[int]] = 8 A__ : Optional[Union[int, Tuple[int]]] = None A__ : int = 12_80 A__ : float = 0.0 A__ : bool = False A__ : jnp.dtype = jnp.floataa A__ : bool = True A__ : int = 0 A__ : bool = False def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : jax.random.KeyArray ) -> FrozenDict: '''simple docstring''' snake_case : Dict = (1, self.in_channels, self.sample_size, self.sample_size) snake_case : Any = jnp.zeros(snake_case__ , dtype=jnp.floataa ) snake_case : List[str] = jnp.ones((1,) , dtype=jnp.intaa ) snake_case : str = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) snake_case , snake_case : Optional[int] = jax.random.split(snake_case__ ) snake_case : Union[str, Any] = {"params": params_rng, "dropout": dropout_rng} return self.init(snake_case__ , snake_case__ , snake_case__ , snake_case__ )["params"] def _SCREAMING_SNAKE_CASE (self : str ) -> Tuple: '''simple docstring''' snake_case : str = self.block_out_channels snake_case : Optional[Any] = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( "At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19." ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. snake_case : Tuple = self.num_attention_heads or self.attention_head_dim # input snake_case : Tuple = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time snake_case : Union[str, Any] = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) snake_case : Dict = FlaxTimestepEmbedding(snake_case__ , dtype=self.dtype ) snake_case : List[str] = self.only_cross_attention if isinstance(snake_case__ , snake_case__ ): snake_case : List[Any] = (only_cross_attention,) * len(self.down_block_types ) if isinstance(snake_case__ , snake_case__ ): snake_case : List[Any] = (num_attention_heads,) * len(self.down_block_types ) # down snake_case : List[Any] = [] snake_case : Optional[int] = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): snake_case : List[Any] = output_channel snake_case : Dict = block_out_channels[i] snake_case : Optional[Any] = i == len(snake_case__ ) - 1 if down_block_type == "CrossAttnDownBlock2D": snake_case : List[Any] = FlaxCrossAttnDownBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: snake_case : Union[str, Any] = FlaxDownBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(snake_case__ ) snake_case : Dict = down_blocks # mid snake_case : Optional[int] = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up snake_case : Optional[Any] = [] snake_case : Optional[int] = list(reversed(snake_case__ ) ) snake_case : Dict = list(reversed(snake_case__ ) ) snake_case : Tuple = list(reversed(snake_case__ ) ) snake_case : Optional[Any] = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): snake_case : Optional[int] = output_channel snake_case : List[Any] = reversed_block_out_channels[i] snake_case : Union[str, Any] = reversed_block_out_channels[min(i + 1 , len(snake_case__ ) - 1 )] snake_case : int = i == len(snake_case__ ) - 1 if up_block_type == "CrossAttnUpBlock2D": snake_case : Any = FlaxCrossAttnUpBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , prev_output_channel=snake_case__ , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: snake_case : Optional[int] = FlaxUpBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , prev_output_channel=snake_case__ , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(snake_case__ ) snake_case : Optional[int] = output_channel snake_case : Tuple = up_blocks # out snake_case : Optional[int] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) snake_case : List[str] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__(self : Dict , snake_case__ : Dict , snake_case__ : Dict , snake_case__ : Optional[int] , snake_case__ : Tuple=None , snake_case__ : Union[str, Any]=None , snake_case__ : bool = True , snake_case__ : bool = False , ) -> Union[FlaxUNetaDConditionOutput, Tuple]: '''simple docstring''' if not isinstance(snake_case__ , jnp.ndarray ): snake_case : List[Any] = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(snake_case__ , jnp.ndarray ) and len(timesteps.shape ) == 0: snake_case : Any = timesteps.astype(dtype=jnp.floataa ) snake_case : int = jnp.expand_dims(snake_case__ , 0 ) snake_case : str = self.time_proj(snake_case__ ) snake_case : str = self.time_embedding(snake_case__ ) # 2. pre-process snake_case : int = jnp.transpose(snake_case__ , (0, 2, 3, 1) ) snake_case : List[Any] = self.conv_in(snake_case__ ) # 3. down snake_case : Optional[int] = (sample,) for down_block in self.down_blocks: if isinstance(snake_case__ , snake_case__ ): snake_case , snake_case : List[Any] = down_block(snake_case__ , snake_case__ , snake_case__ , deterministic=not train ) else: snake_case , snake_case : str = down_block(snake_case__ , snake_case__ , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: snake_case : Tuple = () for down_block_res_sample, down_block_additional_residual in zip( snake_case__ , snake_case__ ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) snake_case : Optional[int] = new_down_block_res_samples # 4. mid snake_case : Optional[int] = self.mid_block(snake_case__ , snake_case__ , snake_case__ , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: snake_case : int = down_block_res_samples[-(self.layers_per_block + 1) :] snake_case : Optional[Any] = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(snake_case__ , snake_case__ ): snake_case : Optional[Any] = up_block( snake_case__ , temb=snake_case__ , encoder_hidden_states=snake_case__ , res_hidden_states_tuple=snake_case__ , deterministic=not train , ) else: snake_case : Dict = up_block(snake_case__ , temb=snake_case__ , res_hidden_states_tuple=snake_case__ , deterministic=not train ) # 6. post-process snake_case : List[str] = self.conv_norm_out(snake_case__ ) snake_case : Any = nn.silu(snake_case__ ) snake_case : Optional[int] = self.conv_out(snake_case__ ) snake_case : Union[str, Any] = jnp.transpose(snake_case__ , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=snake_case__ )
59
0
"""simple docstring""" from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder A: Union[str, Any] = datasets.utils.logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( folder_based_builder.FolderBasedBuilderConfig ): __lowerCAmelCase : bool = None __lowerCAmelCase : bool = None class SCREAMING_SNAKE_CASE__ ( folder_based_builder.FolderBasedBuilder ): __lowerCAmelCase : Union[str, Any] = datasets.Audio() __lowerCAmelCase : Union[str, Any] = 'audio' __lowerCAmelCase : Optional[Any] = AudioFolderConfig __lowerCAmelCase : List[str] # definition at the bottom of the script __lowerCAmelCase : Tuple = AudioClassification(audio_column='audio' , label_column='label' ) A: Union[str, Any] = [ ".aiff", ".au", ".avr", ".caf", ".flac", ".htk", ".svx", ".mat4", ".mat5", ".mpc2k", ".ogg", ".paf", ".pvf", ".raw", ".rf64", ".sd2", ".sds", ".ircam", ".voc", ".w64", ".wav", ".nist", ".wavex", ".wve", ".xi", ".mp3", ".opus", ] A: List[Any] = AUDIO_EXTENSIONS
76
"""simple docstring""" import math import sys def _snake_case ( UpperCamelCase : str ): UpperCAmelCase : Dict = """""" try: with open(UpperCamelCase , """rb""" ) as binary_file: UpperCAmelCase : str = binary_file.read() for dat in data: UpperCAmelCase : List[Any] = F"{dat:08b}" result += curr_byte return result except OSError: print("""File not accessible""" ) sys.exit() def _snake_case ( UpperCamelCase : str ): UpperCAmelCase : Optional[int] = {"""0""": """0""", """1""": """1"""} UpperCAmelCase , UpperCAmelCase : Optional[int] = """""", """""" UpperCAmelCase : int = len(UpperCamelCase ) for i in range(len(UpperCamelCase ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue UpperCAmelCase : Any = lexicon[curr_string] result += last_match_id UpperCAmelCase : Any = last_match_id + """0""" if math.loga(UpperCamelCase ).is_integer(): UpperCAmelCase : Optional[Any] = {} for curr_key in list(UpperCamelCase ): UpperCAmelCase : Dict = lexicon.pop(UpperCamelCase ) UpperCAmelCase : int = new_lex UpperCAmelCase : int = last_match_id + """1""" index += 1 UpperCAmelCase : List[str] = """""" return result def _snake_case ( UpperCamelCase : str , UpperCamelCase : str ): UpperCAmelCase : Dict = 8 try: with open(UpperCamelCase , """wb""" ) as opened_file: UpperCAmelCase : Union[str, Any] = [ to_write[i : i + byte_length] for i in range(0 , len(UpperCamelCase ) , UpperCamelCase ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append("""10000000""" ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(UpperCamelCase , 2 ).to_bytes(1 , byteorder="""big""" ) ) except OSError: print("""File not accessible""" ) sys.exit() def _snake_case ( UpperCamelCase : str ): UpperCAmelCase : Any = 0 for letter in data_bits: if letter == "1": break counter += 1 UpperCAmelCase : List[str] = data_bits[counter:] UpperCAmelCase : Tuple = data_bits[counter + 1 :] return data_bits def _snake_case ( UpperCamelCase : str , UpperCamelCase : str ): UpperCAmelCase : int = read_file_binary(UpperCamelCase ) UpperCAmelCase : str = remove_prefix(UpperCamelCase ) UpperCAmelCase : Any = decompress_data(UpperCamelCase ) write_file_binary(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
76
1
import os import unittest from transformers.models.transfo_xl.tokenization_transfo_xl import VOCAB_FILES_NAMES, TransfoXLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _SCREAMING_SNAKE_CASE ( _a , unittest.TestCase ): snake_case__ : Optional[Any] = TransfoXLTokenizer snake_case__ : List[Any] = False snake_case__ : Tuple = False def _A ( self : str ): super().setUp() UpperCamelCase :Dict = [ """<unk>""", """[CLS]""", """[SEP]""", """want""", """unwanted""", """wa""", """un""", """running""", """,""", """low""", """l""", ] UpperCamelCase :str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _A ( self : List[str] , **__lowerCamelCase : Any ): UpperCamelCase :Any = True return TransfoXLTokenizer.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _A ( self : Any , __lowerCamelCase : int ): UpperCamelCase :List[Any] = """<unk> UNwanted , running""" UpperCamelCase :int = """<unk> unwanted, running""" return input_text, output_text def _A ( self : Tuple ): UpperCamelCase :List[str] = TransfoXLTokenizer(vocab_file=self.vocab_file , lower_case=__lowerCamelCase ) UpperCamelCase :Any = tokenizer.tokenize("""<unk> UNwanted , running""" ) self.assertListEqual(__lowerCamelCase , ["""<unk>""", """unwanted""", """,""", """running"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [0, 4, 8, 7] ) def _A ( self : Optional[Any] ): UpperCamelCase :List[Any] = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) def _A ( self : Union[str, Any] ): UpperCamelCase :int = TransfoXLTokenizer(lower_case=__lowerCamelCase ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo ! how \n Are yoU ? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _A ( self : Tuple ): UpperCamelCase :Any = TransfoXLTokenizer(lower_case=__lowerCamelCase ) UpperCamelCase :Optional[int] = """Hello (bracket) and side-scrolled [and] Henry's $5,000 with 3.34 m. What's up!?""" UpperCamelCase :Optional[int] = [ """Hello""", """(""", """bracket""", """)""", """and""", """side""", """@-@""", """scrolled""", """[""", """and""", """]""", """Henry""", """'s""", """$""", """5""", """@,@""", """000""", """with""", """3""", """@.@""", """34""", """m""", """.""", """What""", """'s""", """up""", """!""", """?""", ] self.assertListEqual(tokenizer.tokenize(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(tokenizer.convert_tokens_to_string(__lowerCamelCase ) , __lowerCamelCase ) def _A ( self : List[Any] ): UpperCamelCase :Any = self.get_tokenizer() UpperCamelCase :List[str] = len(__lowerCamelCase ) tokenizer.add_tokens(["""new1""", """new2"""] ) tokenizer.move_added_token("""new1""" , 1 ) # Check that moved token is not copied (duplicate) self.assertEqual(len(__lowerCamelCase ) , original_len + 2 ) # Check that token is moved to specified id self.assertEqual(tokenizer.encode("""new1""" ) , [1] ) self.assertEqual(tokenizer.decode([1] ) , """new1""" )
38
import re import string import numpy as np import datasets UpperCAmelCase_ : Dict = ''' Returns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list. ''' UpperCAmelCase_ : Any = ''' Args: predictions: List of predicted texts. references: List of reference texts. regexes_to_ignore: List, defaults to None. Regex expressions of characters to ignore when calculating the exact matches. Note: these regexes are removed from the input data before the changes based on the options below (e.g. ignore_case, ignore_punctuation, ignore_numbers) are applied. ignore_case: Boolean, defaults to False. If true, turns everything to lowercase so that capitalization differences are ignored. ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. Returns: exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive. Examples: >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results["exact_match"], 1)) 25.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell"], ignore_case=True, ignore_punctuation=True) >>> print(round(results["exact_match"], 1)) 50.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True) >>> print(round(results["exact_match"], 1)) 75.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["the cat", "theater", "YELLING", "agent007"] >>> preds = ["cat?", "theater", "yelling", "agent"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True) >>> print(round(results["exact_match"], 1)) 100.0 >>> exact_match = datasets.load_metric("exact_match") >>> refs = ["The cat sat on the mat.", "Theaters are great.", "It\'s like comparing oranges and apples."] >>> preds = ["The cat sat on the mat?", "Theaters are great.", "It\'s like comparing apples and oranges."] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results["exact_match"], 1)) 33.3 ''' UpperCAmelCase_ : Tuple = ''' ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _SCREAMING_SNAKE_CASE ( datasets.Metric ): def _A ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , reference_urls=[] , ) def _A ( self : Optional[int] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[str] , __lowerCamelCase : List[str]=None , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : int=False , __lowerCamelCase : Optional[int]=False , ): if regexes_to_ignore is not None: for s in regexes_to_ignore: UpperCamelCase :str = np.array([re.sub(__lowerCamelCase , """""" , __lowerCamelCase ) for x in predictions] ) UpperCamelCase :Tuple = np.array([re.sub(__lowerCamelCase , """""" , __lowerCamelCase ) for x in references] ) else: UpperCamelCase :Any = np.asarray(__lowerCamelCase ) UpperCamelCase :str = np.asarray(__lowerCamelCase ) if ignore_case: UpperCamelCase :Tuple = np.char.lower(__lowerCamelCase ) UpperCamelCase :Any = np.char.lower(__lowerCamelCase ) if ignore_punctuation: UpperCamelCase :Optional[int] = string.punctuation.maketrans("""""" , """""" , string.punctuation ) UpperCamelCase :Optional[Any] = np.char.translate(__lowerCamelCase , table=__lowerCamelCase ) UpperCamelCase :List[str] = np.char.translate(__lowerCamelCase , table=__lowerCamelCase ) if ignore_numbers: UpperCamelCase :Tuple = string.digits.maketrans("""""" , """""" , string.digits ) UpperCamelCase :Dict = np.char.translate(__lowerCamelCase , table=__lowerCamelCase ) UpperCamelCase :Tuple = np.char.translate(__lowerCamelCase , table=__lowerCamelCase ) UpperCamelCase :int = predictions == references return {"exact_match": np.mean(__lowerCamelCase ) * 100}
38
1
'''simple docstring''' import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" lowerCamelCase = MODEL_FOR_MASKED_LM_MAPPING lowerCamelCase = TF_MODEL_FOR_MASKED_LM_MAPPING def UpperCAmelCase_ ( self ) -> List[str]: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[int] = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , top_k=2 , framework="""tf""" ) A_ : Any = unmasker("""My name is <mask>""" ) self.assertEqual( nested_simplify(_lowerCamelCase , decimals=6 ) , [ {"""sequence""": """My name is grouped""", """score""": 2.1e-05, """token""": 3_8015, """token_str""": """ grouped"""}, {"""sequence""": """My name is accuser""", """score""": 2.1e-05, """token""": 2_5506, """token_str""": """ accuser"""}, ] , ) A_ : List[str] = unmasker("""The largest city in France is <mask>""" ) self.assertEqual( nested_simplify(_lowerCamelCase , decimals=6 ) , [ { """sequence""": """The largest city in France is grouped""", """score""": 2.1e-05, """token""": 3_8015, """token_str""": """ grouped""", }, { """sequence""": """The largest city in France is accuser""", """score""": 2.1e-05, """token""": 2_5506, """token_str""": """ accuser""", }, ] , ) A_ : Optional[Any] = unmasker("""My name is <mask>""" , targets=[""" Patrick""", """ Clara""", """ Teven"""] , top_k=3 ) self.assertEqual( nested_simplify(_lowerCamelCase , decimals=6 ) , [ {"""sequence""": """My name is Clara""", """score""": 2e-05, """token""": 1_3606, """token_str""": """ Clara"""}, {"""sequence""": """My name is Patrick""", """score""": 2e-05, """token""": 3499, """token_str""": """ Patrick"""}, {"""sequence""": """My name is Te""", """score""": 1.9e-05, """token""": 2941, """token_str""": """ Te"""}, ] , ) @require_torch def UpperCAmelCase_ ( self ) -> Any: A_ : List[Any] = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , top_k=2 , framework="""pt""" ) A_ : Optional[int] = unmasker("""My name is <mask>""" ) self.assertEqual( nested_simplify(_lowerCamelCase , decimals=6 ) , [ {"""sequence""": """My name is Maul""", """score""": 2.2e-05, """token""": 3_5676, """token_str""": """ Maul"""}, {"""sequence""": """My name isELS""", """score""": 2.2e-05, """token""": 1_6416, """token_str""": """ELS"""}, ] , ) A_ : Dict = unmasker("""The largest city in France is <mask>""" ) self.assertEqual( nested_simplify(_lowerCamelCase , decimals=6 ) , [ { """sequence""": """The largest city in France is Maul""", """score""": 2.2e-05, """token""": 3_5676, """token_str""": """ Maul""", }, {"""sequence""": """The largest city in France isELS""", """score""": 2.2e-05, """token""": 1_6416, """token_str""": """ELS"""}, ] , ) A_ : int = unmasker("""My name is <mask>""" , targets=[""" Patrick""", """ Clara""", """ Teven"""] , top_k=3 ) self.assertEqual( nested_simplify(_lowerCamelCase , decimals=6 ) , [ {"""sequence""": """My name is Patrick""", """score""": 2.1e-05, """token""": 3499, """token_str""": """ Patrick"""}, {"""sequence""": """My name is Te""", """score""": 2e-05, """token""": 2941, """token_str""": """ Te"""}, {"""sequence""": """My name is Clara""", """score""": 2e-05, """token""": 1_3606, """token_str""": """ Clara"""}, ] , ) A_ : Optional[Any] = unmasker("""My name is <mask> <mask>""" , top_k=2 ) self.assertEqual( nested_simplify(_lowerCamelCase , decimals=6 ) , [ [ { """score""": 2.2e-05, """token""": 3_5676, """token_str""": """ Maul""", """sequence""": """<s>My name is Maul<mask></s>""", }, {"""score""": 2.2e-05, """token""": 1_6416, """token_str""": """ELS""", """sequence""": """<s>My name isELS<mask></s>"""}, ], [ { """score""": 2.2e-05, """token""": 3_5676, """token_str""": """ Maul""", """sequence""": """<s>My name is<mask> Maul</s>""", }, {"""score""": 2.2e-05, """token""": 1_6416, """token_str""": """ELS""", """sequence""": """<s>My name is<mask>ELS</s>"""}, ], ] , ) @require_torch_gpu def UpperCAmelCase_ ( self ) -> str: A_ : Any = pipeline("""fill-mask""" , model="""hf-internal-testing/tiny-random-distilbert""" , device=0 , framework="""pt""" ) # convert model to fp16 pipe.model.half() A_ : List[str] = pipe("""Paris is the [MASK] of France.""" ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(_lowerCamelCase , _lowerCamelCase ) @slow @require_torch def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : str = pipeline(task="""fill-mask""" , model="""distilroberta-base""" , top_k=2 , framework="""pt""" ) self.run_large_test(_lowerCamelCase ) @slow @require_tf def UpperCAmelCase_ ( self ) -> List[str]: A_ : Union[str, Any] = pipeline(task="""fill-mask""" , model="""distilroberta-base""" , top_k=2 , framework="""tf""" ) self.run_large_test(_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase ) -> str: A_ : Tuple = unmasker("""My name is <mask>""" ) self.assertEqual( nested_simplify(_lowerCamelCase ) , [ {"""sequence""": """My name is John""", """score""": 0.008, """token""": 610, """token_str""": """ John"""}, {"""sequence""": """My name is Chris""", """score""": 0.007, """token""": 1573, """token_str""": """ Chris"""}, ] , ) A_ : str = unmasker("""The largest city in France is <mask>""" ) self.assertEqual( nested_simplify(_lowerCamelCase ) , [ { """sequence""": """The largest city in France is Paris""", """score""": 0.251, """token""": 2201, """token_str""": """ Paris""", }, { """sequence""": """The largest city in France is Lyon""", """score""": 0.214, """token""": 1_2790, """token_str""": """ Lyon""", }, ] , ) A_ : Union[str, Any] = unmasker("""My name is <mask>""" , targets=[""" Patrick""", """ Clara""", """ Teven"""] , top_k=3 ) self.assertEqual( nested_simplify(_lowerCamelCase ) , [ {"""sequence""": """My name is Patrick""", """score""": 0.005, """token""": 3499, """token_str""": """ Patrick"""}, {"""sequence""": """My name is Clara""", """score""": 0.000, """token""": 1_3606, """token_str""": """ Clara"""}, {"""sequence""": """My name is Te""", """score""": 0.000, """token""": 2941, """token_str""": """ Te"""}, ] , ) @require_torch def UpperCAmelCase_ ( self ) -> Union[str, Any]: A_ : str = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , framework="""pt""" ) A_ : int = None A_ : List[Any] = None self.run_pipeline_test(_lowerCamelCase , [] ) @require_tf def UpperCAmelCase_ ( self ) -> Tuple: A_ : List[Any] = pipeline(task="""fill-mask""" , model="""sshleifer/tiny-distilroberta-base""" , framework="""tf""" ) A_ : Dict = None A_ : Any = None self.run_pipeline_test(_lowerCamelCase , [] ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> str: if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest("""The provided tokenizer has no mask token, (probably reformer or wav2vec2)""" ) A_ : str = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase ) A_ : Any = [ F"This is another {tokenizer.mask_token} test", ] return fill_masker, examples def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> Any: A_ : Dict = fill_masker.tokenizer A_ : List[Any] = fill_masker.model A_ : Optional[Any] = fill_masker( F"This is a {tokenizer.mask_token}" , ) self.assertEqual( _lowerCamelCase , [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ] , ) A_ : Any = fill_masker([F"This is a {tokenizer.mask_token}"] ) self.assertEqual( _lowerCamelCase , [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ] , ) A_ : Union[str, Any] = fill_masker([F"This is a {tokenizer.mask_token}", F"Another {tokenizer.mask_token} great test."] ) self.assertEqual( _lowerCamelCase , [ [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ], [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ], ] , ) with self.assertRaises(_lowerCamelCase ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(_lowerCamelCase ): fill_masker("""This is""" ) self.run_test_top_k(_lowerCamelCase , _lowerCamelCase ) self.run_test_targets(_lowerCamelCase , _lowerCamelCase ) self.run_test_top_k_targets(_lowerCamelCase , _lowerCamelCase ) self.fill_mask_with_duplicate_targets_and_top_k(_lowerCamelCase , _lowerCamelCase ) self.fill_mask_with_multiple_masks(_lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> Optional[Any]: A_ : Tuple = tokenizer.get_vocab() A_ : List[Any] = sorted(vocab.keys() )[:2] # Pipeline argument A_ : List[Any] = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase , targets=_lowerCamelCase ) A_ : Optional[int] = fill_masker(F"This is a {tokenizer.mask_token}" ) self.assertEqual( _lowerCamelCase , [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ] , ) A_ : Any = {vocab[el] for el in targets} self.assertEqual({el["""token"""] for el in outputs} , _lowerCamelCase ) A_ : List[str] = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["""token_str"""] for el in outputs} , set(_lowerCamelCase ) ) # Call argument A_ : Optional[Any] = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase ) A_ : Any = fill_masker(F"This is a {tokenizer.mask_token}" , targets=_lowerCamelCase ) self.assertEqual( _lowerCamelCase , [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ] , ) A_ : int = {vocab[el] for el in targets} self.assertEqual({el["""token"""] for el in outputs} , _lowerCamelCase ) A_ : Any = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el["""token_str"""] for el in outputs} , set(_lowerCamelCase ) ) # Score equivalence A_ : Tuple = fill_masker(F"This is a {tokenizer.mask_token}" , targets=_lowerCamelCase ) A_ : Optional[Any] = [top_mask["""token_str"""] for top_mask in outputs] A_ : List[Any] = [top_mask["""score"""] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(_lowerCamelCase ) == set(_lowerCamelCase ): A_ : Dict = fill_masker(F"This is a {tokenizer.mask_token}" , targets=_lowerCamelCase ) A_ : Any = [top_mask["""score"""] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(_lowerCamelCase ) , nested_simplify(_lowerCamelCase ) ) # Raises with invalid with self.assertRaises(_lowerCamelCase ): A_ : str = fill_masker(F"This is a {tokenizer.mask_token}" , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(_lowerCamelCase ): A_ : int = fill_masker(F"This is a {tokenizer.mask_token}" , targets=[""""""] ) with self.assertRaises(_lowerCamelCase ): A_ : Any = fill_masker(F"This is a {tokenizer.mask_token}" , targets="""""" ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> str: A_ : Optional[int] = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase , top_k=2 ) A_ : Any = fill_masker(F"This is a {tokenizer.mask_token}" ) self.assertEqual( _lowerCamelCase , [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ] , ) A_ : Union[str, Any] = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase ) A_ : List[Any] = fill_masker(F"This is a {tokenizer.mask_token}" , top_k=2 ) self.assertEqual( _lowerCamelCase , [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ] , ) self.assertEqual(nested_simplify(_lowerCamelCase ) , nested_simplify(_lowerCamelCase ) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[Any] = tokenizer.get_vocab() A_ : int = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase ) # top_k=2, ntargets=3 A_ : int = sorted(vocab.keys() )[:3] A_ : List[Any] = fill_masker(F"This is a {tokenizer.mask_token}" , top_k=2 , targets=_lowerCamelCase ) # If we use the most probably targets, and filter differently, we should still # have the same results A_ : List[Any] = [el["""token_str"""] for el in sorted(_lowerCamelCase , key=lambda _lowerCamelCase : x["score"] , reverse=_lowerCamelCase )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(_lowerCamelCase ).issubset(_lowerCamelCase ): A_ : Union[str, Any] = fill_masker(F"This is a {tokenizer.mask_token}" , top_k=3 , targets=_lowerCamelCase ) # They should yield exactly the same result self.assertEqual(nested_simplify(_lowerCamelCase ) , nested_simplify(_lowerCamelCase ) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> str: A_ : Any = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase ) A_ : List[str] = tokenizer.get_vocab() # String duplicates + id duplicates A_ : List[str] = sorted(vocab.keys() )[:3] A_ : List[str] = [targets[0], targets[1], targets[0], targets[2], targets[1]] A_ : Optional[int] = fill_masker(F"My name is {tokenizer.mask_token}" , targets=_lowerCamelCase , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(_lowerCamelCase ) , 3 ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> str: A_ : Any = FillMaskPipeline(model=_lowerCamelCase , tokenizer=_lowerCamelCase ) A_ : Optional[int] = fill_masker( F"This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}" , top_k=2 ) self.assertEqual( _lowerCamelCase , [ [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ], [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ], [ {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, {"""sequence""": ANY(_lowerCamelCase ), """score""": ANY(_lowerCamelCase ), """token""": ANY(_lowerCamelCase ), """token_str""": ANY(_lowerCamelCase )}, ], ] , )
164
'''simple docstring''' import math def UpperCAmelCase ( a_ ) -> bool: """simple docstring""" return math.sqrt(a_ ) * math.sqrt(a_ ) == num def UpperCAmelCase ( a_ ) -> bool: """simple docstring""" A_ : Tuple = 0 A_ : Tuple = n while left <= right: A_ : int = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: A_ : Dict = mid - 1 else: A_ : str = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
164
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) def UpperCamelCase_( snake_case : Union[str, Any] , snake_case : Optional[Any]=False ): '''simple docstring''' snake_case_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" snake_case_ = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def UpperCamelCase_( snake_case : Any , snake_case : Tuple , snake_case : str=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: snake_case_ = "" else: snake_case_ = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case_ = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) snake_case_ = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict snake_case_ = in_proj_weight[ : config.hidden_size, : ] snake_case_ = in_proj_bias[: config.hidden_size] snake_case_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case_ = in_proj_weight[ -config.hidden_size :, : ] snake_case_ = in_proj_bias[-config.hidden_size :] def UpperCamelCase_( snake_case : Tuple ): '''simple docstring''' snake_case_ = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(snake_case , snake_case ) def UpperCamelCase_( snake_case : List[str] , snake_case : Union[str, Any] , snake_case : int ): '''simple docstring''' snake_case_ = dct.pop(snake_case ) snake_case_ = val def UpperCamelCase_( ): '''simple docstring''' snake_case_ = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ = Image.open(requests.get(snake_case , stream=snake_case ).raw ) return im @torch.no_grad() def UpperCamelCase_( snake_case : Optional[Any] , snake_case : Any , snake_case : str=True ): '''simple docstring''' snake_case_ = ViTConfig() # patch_size if model_name[-1] == "8": snake_case_ = 8 # set labels if required if not base_model: snake_case_ = 1_0_0_0 snake_case_ = "huggingface/label-files" snake_case_ = "imagenet-1k-id2label.json" snake_case_ = json.load(open(hf_hub_download(snake_case , snake_case , repo_type="dataset" ) , "r" ) ) snake_case_ = {int(snake_case ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: snake_case_ = 3_8_4 snake_case_ = 1_5_3_6 snake_case_ = 1_2 snake_case_ = 6 # load original model from torch hub snake_case_ = torch.hub.load("facebookresearch/dino:main" , snake_case ) original_model.eval() # load state_dict of original model, remove and rename some keys snake_case_ = original_model.state_dict() if base_model: remove_classification_head_(snake_case ) snake_case_ = create_rename_keys(snake_case , base_model=snake_case ) for src, dest in rename_keys: rename_key(snake_case , snake_case , snake_case ) read_in_q_k_v(snake_case , snake_case , snake_case ) # load HuggingFace model if base_model: snake_case_ = ViTModel(snake_case , add_pooling_layer=snake_case ).eval() else: snake_case_ = ViTForImageClassification(snake_case ).eval() model.load_state_dict(snake_case ) # Check outputs on an image, prepared by ViTImageProcessor snake_case_ = ViTImageProcessor() snake_case_ = image_processor(images=prepare_img() , return_tensors="pt" ) snake_case_ = encoding["pixel_values"] snake_case_ = model(snake_case ) if base_model: snake_case_ = original_model(snake_case ) assert torch.allclose(snake_case , outputs.last_hidden_state[:, 0, :] , atol=1e-1 ) else: snake_case_ = original_model(snake_case ) assert logits.shape == outputs.logits.shape assert torch.allclose(snake_case , outputs.logits , atol=1e-3 ) Path(snake_case ).mkdir(exist_ok=snake_case ) print(f'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(snake_case ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(snake_case ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="dino_vitb16", type=str, help="Name of the model trained with DINO you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--base_model", action="store_true", help="Whether to only convert the base model (no projection head weights).", ) parser.set_defaults(base_model=True) _SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
85
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer a_ : Dict = logging.get_logger(__name__) a_ : Optional[Any] = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} # See all BART models at https://huggingface.co/models?filter=bart a_ : int = { "vocab_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/vocab.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/vocab.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json", }, "merges_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/merges.txt", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/merges.txt", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt", }, "tokenizer_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json", }, } a_ : int = { "facebook/bart-base": 1_0_2_4, "facebook/bart-large": 1_0_2_4, "facebook/bart-large-mnli": 1_0_2_4, "facebook/bart-large-cnn": 1_0_2_4, "facebook/bart-large-xsum": 1_0_2_4, "yjernite/bart_eli5": 1_0_2_4, } class a ( _SCREAMING_SNAKE_CASE ): _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase = ["""input_ids""", """attention_mask"""] _lowerCAmelCase = BartTokenizer def __init__( self , __magic_name__=None , __magic_name__=None , __magic_name__=None , __magic_name__="replace" , __magic_name__="<s>" , __magic_name__="</s>" , __magic_name__="</s>" , __magic_name__="<s>" , __magic_name__="<unk>" , __magic_name__="<pad>" , __magic_name__="<mask>" , __magic_name__=False , __magic_name__=True , **__magic_name__ , ) -> List[Any]: super().__init__( __magic_name__ , __magic_name__ , tokenizer_file=__magic_name__ , errors=__magic_name__ , bos_token=__magic_name__ , eos_token=__magic_name__ , sep_token=__magic_name__ , cls_token=__magic_name__ , unk_token=__magic_name__ , pad_token=__magic_name__ , mask_token=__magic_name__ , add_prefix_space=__magic_name__ , trim_offsets=__magic_name__ , **__magic_name__ , ) _a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , __magic_name__ ) != add_prefix_space: _a = getattr(__magic_name__ , pre_tok_state.pop('type' ) ) _a = add_prefix_space _a = pre_tok_class(**__magic_name__ ) _a = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` _a = 'post_processor' _a = getattr(self.backend_tokenizer , __magic_name__ , __magic_name__ ) if tokenizer_component_instance: _a = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: _a = tuple(state['sep'] ) if "cls" in state: _a = tuple(state['cls'] ) _a = False if state.get('add_prefix_space' , __magic_name__ ) != add_prefix_space: _a = add_prefix_space _a = True if state.get('trim_offsets' , __magic_name__ ) != trim_offsets: _a = trim_offsets _a = True if changes_to_apply: _a = getattr(__magic_name__ , state.pop('type' ) ) _a = component_class(**__magic_name__ ) setattr(self.backend_tokenizer , __magic_name__ , __magic_name__ ) @property def __UpperCAmelCase ( self ) -> str: if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def __UpperCAmelCase ( self , __magic_name__ ) -> Union[str, Any]: _a = AddedToken(__magic_name__ , lstrip=__magic_name__ , rstrip=__magic_name__ ) if isinstance(__magic_name__ , __magic_name__ ) else value _a = value def __UpperCAmelCase ( self , *__magic_name__ , **__magic_name__ ) -> BatchEncoding: _a = kwargs.get('is_split_into_words' , __magic_name__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*__magic_name__ , **__magic_name__ ) def __UpperCAmelCase ( self , *__magic_name__ , **__magic_name__ ) -> BatchEncoding: _a = kwargs.get('is_split_into_words' , __magic_name__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' 'to use it with pretokenized inputs.' ) return super()._encode_plus(*__magic_name__ , **__magic_name__ ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = None ) -> Tuple[str]: _a = self._tokenizer.model.save(__magic_name__ , name=__magic_name__ ) return tuple(__magic_name__ ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__=None ) -> Any: _a = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = None ) -> List[int]: _a = [self.sep_token_id] _a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
168
0
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase__: def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase=12 ,__UpperCAmelCase=7 ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=99 ,__UpperCAmelCase=32 ,__UpperCAmelCase=32 ,__UpperCAmelCase=2 ,__UpperCAmelCase=4 ,__UpperCAmelCase=37 ,__UpperCAmelCase=0.1 ,__UpperCAmelCase=0.1 ,__UpperCAmelCase=5_12 ,__UpperCAmelCase=0.0_2 ,__UpperCAmelCase=0 ,__UpperCAmelCase=None ,) -> int: A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = projection_dim A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = dropout A__ = attention_dropout A__ = max_position_embeddings A__ = initializer_range A__ = scope A__ = bos_token_id def snake_case__ ( self ) -> Optional[Any]: A__ = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: A__ = input_mask.numpy() A__ , A__ = input_mask.shape A__ = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(__UpperCAmelCase ): A__ = 1 A__ = 0 A__ = self.get_config() return config, input_ids, tf.convert_to_tensor(__UpperCAmelCase ) def snake_case__ ( self ) -> List[Any]: return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) -> List[str]: A__ = TFBlipTextModel(config=__UpperCAmelCase ) A__ = model(__UpperCAmelCase ,attention_mask=__UpperCAmelCase ,training=__UpperCAmelCase ) A__ = model(__UpperCAmelCase ,training=__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def snake_case__ ( self ) -> List[str]: A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCamelCase__( __A , unittest.TestCase ): lowerCAmelCase__ : List[str] = (TFBlipTextModel,) if is_tf_available() else () lowerCAmelCase__ : List[Any] = False lowerCAmelCase__ : Dict = False lowerCAmelCase__ : List[Any] = False def snake_case__ ( self ) -> Optional[int]: A__ = BlipTextModelTester(self ) A__ = ConfigTester(self ,config_class=__UpperCAmelCase ,hidden_size=37 ) def snake_case__ ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() def snake_case__ ( self ) -> Union[str, Any]: A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def snake_case__ ( self ) -> List[str]: pass def snake_case__ ( self ) -> Union[str, Any]: pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def snake_case__ ( self ) -> str: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def snake_case__ ( self ) -> Tuple: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def snake_case__ ( self ) -> Dict: pass @slow def snake_case__ ( self ) -> str: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = TFBlipTextModel.from_pretrained(__UpperCAmelCase ) self.assertIsNotNone(__UpperCAmelCase ) def snake_case__ ( self ,__UpperCAmelCase=True ) -> List[Any]: super().test_pt_tf_model_equivalence(allow_missing_keys=__UpperCAmelCase )
354
"""simple docstring""" import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Accelerator def UpperCAmelCase ( UpperCamelCase__ ): """simple docstring""" A__ = fname.split(os.path.sep )[-1] return re.search(r'^(.*)_\d+\.jpg$' , UpperCamelCase__ ).groups()[0] class UpperCamelCase__( __A ): def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase=None ,__UpperCAmelCase=None ) -> List[str]: A__ = file_names A__ = image_transform A__ = label_to_id def __len__( self ) -> Dict: return len(self.file_names ) def __getitem__( self ,__UpperCAmelCase ) -> Union[str, Any]: A__ = self.file_names[idx] A__ = PIL.Image.open(__UpperCAmelCase ) A__ = raw_image.convert('RGB' ) if self.image_transform is not None: A__ = self.image_transform(__UpperCAmelCase ) A__ = extract_label(__UpperCAmelCase ) if self.label_to_id is not None: A__ = self.label_to_id[label] return {"image": image, "label": label} def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" if args.with_tracking: A__ = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with='all' , project_dir=args.project_dir ) else: A__ = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs A__ = config['lr'] A__ = int(config['num_epochs'] ) A__ = int(config['seed'] ) A__ = int(config['batch_size'] ) A__ = config['image_size'] if not isinstance(UpperCamelCase__ , (list, tuple) ): A__ = (image_size, image_size) # Parse out whether we are saving every epoch or after a certain number of batches if hasattr(args.checkpointing_steps , 'isdigit' ): if args.checkpointing_steps == "epoch": A__ = args.checkpointing_steps elif args.checkpointing_steps.isdigit(): A__ = int(args.checkpointing_steps ) else: raise ValueError( F'''Argument `checkpointing_steps` must be either a number or `epoch`. `{args.checkpointing_steps}` passed.''' ) else: A__ = None # We need to initialize the trackers we use, and also store our configuration if args.with_tracking: A__ = os.path.split(UpperCamelCase__ )[-1].split('.' )[0] accelerator.init_trackers(UpperCamelCase__ , UpperCamelCase__ ) # Grab all the image filenames A__ = [os.path.join(args.data_dir , UpperCamelCase__ ) for fname in os.listdir(args.data_dir ) if fname.endswith('.jpg' )] # Build the label correspondences A__ = [extract_label(UpperCamelCase__ ) for fname in file_names] A__ = list(set(UpperCamelCase__ ) ) id_to_label.sort() A__ = {lbl: i for i, lbl in enumerate(UpperCamelCase__ )} # Set the seed before splitting the data. np.random.seed(UpperCamelCase__ ) torch.manual_seed(UpperCamelCase__ ) torch.cuda.manual_seed_all(UpperCamelCase__ ) # Split our filenames between train and validation A__ = np.random.permutation(len(UpperCamelCase__ ) ) A__ = int(0.8 * len(UpperCamelCase__ ) ) A__ = random_perm[:cut] A__ = random_perm[cut:] # For training we use a simple RandomResizedCrop A__ = Compose([RandomResizedCrop(UpperCamelCase__ , scale=(0.5, 1.0) ), ToTensor()] ) A__ = PetsDataset( [file_names[i] for i in train_split] , image_transform=UpperCamelCase__ , label_to_id=UpperCamelCase__ ) # For evaluation, we use a deterministic Resize A__ = Compose([Resize(UpperCamelCase__ ), ToTensor()] ) A__ = PetsDataset([file_names[i] for i in eval_split] , image_transform=UpperCamelCase__ , label_to_id=UpperCamelCase__ ) # Instantiate dataloaders. A__ = DataLoader(UpperCamelCase__ , shuffle=UpperCamelCase__ , batch_size=UpperCamelCase__ , num_workers=4 ) A__ = DataLoader(UpperCamelCase__ , shuffle=UpperCamelCase__ , batch_size=UpperCamelCase__ , num_workers=4 ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) A__ = create_model('resnet50d' , pretrained=UpperCamelCase__ , num_classes=len(UpperCamelCase__ ) ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). A__ = model.to(accelerator.device ) # Freezing the base model for param in model.parameters(): A__ = False for param in model.get_classifier().parameters(): A__ = True # We normalize the batches of images to be a bit faster. A__ = torch.tensor(model.default_cfg['mean'] )[None, :, None, None].to(accelerator.device ) A__ = torch.tensor(model.default_cfg['std'] )[None, :, None, None].to(accelerator.device ) # Instantiate optimizer A__ = torch.optim.Adam(params=model.parameters() , lr=lr / 25 ) # Instantiate learning rate scheduler A__ = OneCycleLR(optimizer=UpperCamelCase__ , max_lr=UpperCamelCase__ , epochs=UpperCamelCase__ , steps_per_epoch=len(UpperCamelCase__ ) ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. A__ , A__ , A__ , A__ , A__ = accelerator.prepare( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # We need to keep track of how many total steps we have iterated over A__ = 0 # We also need to keep track of the starting epoch so files are named properly A__ = 0 # Potentially load in the weights and states from a previous save if args.resume_from_checkpoint: if args.resume_from_checkpoint is not None or args.resume_from_checkpoint != "": accelerator.print(F'''Resumed from checkpoint: {args.resume_from_checkpoint}''' ) accelerator.load_state(args.resume_from_checkpoint ) A__ = os.path.basename(args.resume_from_checkpoint ) else: # Get the most recent checkpoint A__ = [f.name for f in os.scandir(os.getcwd() ) if f.is_dir()] dirs.sort(key=os.path.getctime ) A__ = dirs[-1] # Sorts folders by date modified, most recent checkpoint is the last # Extract `epoch_{i}` or `step_{i}` A__ = os.path.splitext(UpperCamelCase__ )[0] if "epoch" in training_difference: A__ = int(training_difference.replace('epoch_' , '' ) ) + 1 A__ = None else: A__ = int(training_difference.replace('step_' , '' ) ) A__ = resume_step // len(UpperCamelCase__ ) resume_step -= starting_epoch * len(UpperCamelCase__ ) # Now we train the model for epoch in range(UpperCamelCase__ , UpperCamelCase__ ): model.train() if args.with_tracking: A__ = 0 if args.resume_from_checkpoint and epoch == starting_epoch and resume_step is not None: # We need to skip steps until we reach the resumed step A__ = accelerator.skip_first_batches(UpperCamelCase__ , UpperCamelCase__ ) overall_step += resume_step else: # After the first iteration though, we need to go back to the original dataloader A__ = train_dataloader for batch in active_dataloader: # We could avoid this line since we set the accelerator with `device_placement=True`. A__ = {k: v.to(accelerator.device ) for k, v in batch.items()} A__ = (batch['image'] - mean) / std A__ = model(UpperCamelCase__ ) A__ = torch.nn.functional.cross_entropy(UpperCamelCase__ , batch['label'] ) # We keep track of the loss at each epoch if args.with_tracking: total_loss += loss.detach().float() accelerator.backward(UpperCamelCase__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A__ = F'''step_{overall_step}''' if overall_step % checkpointing_steps == 0: if args.output_dir is not None: A__ = os.path.join(args.output_dir , UpperCamelCase__ ) accelerator.save_state(UpperCamelCase__ ) model.eval() A__ = 0 A__ = 0 for step, batch in enumerate(UpperCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. A__ = {k: v.to(accelerator.device ) for k, v in batch.items()} A__ = (batch['image'] - mean) / std with torch.no_grad(): A__ = model(UpperCamelCase__ ) A__ = outputs.argmax(dim=-1 ) A__ , A__ = accelerator.gather_for_metrics((predictions, batch['label']) ) A__ = predictions == references num_elems += accurate_preds.shape[0] accurate += accurate_preds.long().sum() A__ = accurate.item() / num_elems # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}: {100 * eval_metric:.2f}''' ) if args.with_tracking: accelerator.log( { 'accuracy': 100 * eval_metric, 'train_loss': total_loss.item() / len(UpperCamelCase__ ), 'epoch': epoch, } , step=UpperCamelCase__ , ) if checkpointing_steps == "epoch": A__ = F'''epoch_{epoch}''' if args.output_dir is not None: A__ = os.path.join(args.output_dir , UpperCamelCase__ ) accelerator.save_state(UpperCamelCase__ ) if args.with_tracking: accelerator.end_training() def UpperCAmelCase ( ): """simple docstring""" A__ = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument('--data_dir' , required=UpperCamelCase__ , help='The data folder on disk.' ) parser.add_argument('--fp16' , action='store_true' , help='If passed, will use FP16 training.' ) parser.add_argument( '--mixed_precision' , type=UpperCamelCase__ , default=UpperCamelCase__ , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) parser.add_argument( '--checkpointing_steps' , type=UpperCamelCase__ , default=UpperCamelCase__ , help='Whether the various states should be saved at the end of every n steps, or \'epoch\' for each epoch.' , ) parser.add_argument( '--output_dir' , type=UpperCamelCase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--resume_from_checkpoint' , type=UpperCamelCase__ , default=UpperCamelCase__ , help='If the training should continue from a checkpoint folder.' , ) parser.add_argument( '--with_tracking' , action='store_true' , help='Whether to load in all available experiment trackers from the environment and use them for logging.' , ) parser.add_argument( '--project_dir' , type=UpperCamelCase__ , default='logs' , help='Location on where to store experiment tracking logs` and relevent project information' , ) A__ = parser.parse_args() A__ = {'lr': 3E-2, 'num_epochs': 3, 'seed': 42, 'batch_size': 64, 'image_size': 224} training_function(UpperCamelCase__ , UpperCamelCase__ ) if __name__ == "__main__": main()
154
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : int , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int]=7 , _lowerCAmelCase : Union[str, Any]=3 , _lowerCAmelCase : str=1_0 , _lowerCAmelCase : List[Any]=1_8 , _lowerCAmelCase : Tuple=3_0 , _lowerCAmelCase : Optional[int]=4_0_0 , _lowerCAmelCase : Optional[Any]=True , _lowerCAmelCase : Union[str, Any]=None , _lowerCAmelCase : Dict=True , _lowerCAmelCase : int=[0.5, 0.5, 0.5] , _lowerCAmelCase : Optional[Any]=[0.5, 0.5, 0.5] , _lowerCAmelCase : Dict=None , ) -> Any: """simple docstring""" snake_case_ = size if size is not None else {"shortest_edge": 1_8} snake_case_ = crop_size if crop_size is not None else {"height": 1_8, "width": 1_8} snake_case_ = parent snake_case_ = batch_size snake_case_ = num_channels snake_case_ = num_frames snake_case_ = image_size snake_case_ = min_resolution snake_case_ = max_resolution snake_case_ = do_resize snake_case_ = size snake_case_ = do_normalize snake_case_ = image_mean snake_case_ = image_std snake_case_ = crop_size def lowerCAmelCase__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __lowerCAmelCase ( a , unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = VivitImageProcessor if is_vision_available() else None def lowerCAmelCase__ ( self : Optional[int] ) -> Dict: """simple docstring""" snake_case_ = VivitImageProcessingTester(self ) @property def lowerCAmelCase__ ( self : Any ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , "image_mean" ) ) self.assertTrue(hasattr(_lowerCAmelCase , "image_std" ) ) self.assertTrue(hasattr(_lowerCAmelCase , "do_normalize" ) ) self.assertTrue(hasattr(_lowerCAmelCase , "do_resize" ) ) self.assertTrue(hasattr(_lowerCAmelCase , "do_center_crop" ) ) self.assertTrue(hasattr(_lowerCAmelCase , "size" ) ) def lowerCAmelCase__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 1_8} ) self.assertEqual(image_processor.crop_size , {"height": 1_8, "width": 1_8} ) snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {"shortest_edge": 4_2} ) self.assertEqual(image_processor.crop_size , {"height": 8_4, "width": 8_4} ) def lowerCAmelCase__ ( self : Dict ) -> Optional[Any]: """simple docstring""" # Initialize image_processing snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos snake_case_ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase ) for video in video_inputs: self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input snake_case_ = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched snake_case_ = image_processing(_lowerCAmelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def lowerCAmelCase__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" # Initialize image_processing snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , numpify=_lowerCAmelCase ) for video in video_inputs: self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input snake_case_ = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched snake_case_ = image_processing(_lowerCAmelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def lowerCAmelCase__ ( self : Dict ) -> List[Any]: """simple docstring""" # Initialize image_processing snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , torchify=_lowerCAmelCase ) for video in video_inputs: self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input snake_case_ = image_processing(video_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched snake_case_ = image_processing(_lowerCAmelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
159
from __future__ import annotations from collections import namedtuple def _lowerCAmelCase ( lowerCAmelCase_ :float , lowerCAmelCase_ :float , lowerCAmelCase_ :float )->tuple: '''simple docstring''' snake_case_ = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
159
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase _UpperCamelCase : Dict = logging.get_logger(__name__) _UpperCamelCase : Dict = { 'allenai/longformer-base-4096': 'https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json', 'allenai/longformer-large-4096': 'https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json', 'allenai/longformer-large-4096-finetuned-triviaqa': ( 'https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json' ), 'allenai/longformer-base-4096-extra.pos.embd.only': ( 'https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json' ), 'allenai/longformer-large-4096-extra.pos.embd.only': ( 'https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json' ), } class snake_case ( UpperCAmelCase ): __magic_name__ = '''longformer''' def __init__( self : Optional[Any] , A : Union[List[int], int] = 5_1_2 , A : int = 2 , A : int = 1 , A : int = 0 , A : int = 2 , A : int = 3_0_5_2_2 , A : int = 7_6_8 , A : int = 1_2 , A : int = 1_2 , A : int = 3_0_7_2 , A : str = "gelu" , A : float = 0.1 , A : float = 0.1 , A : int = 5_1_2 , A : int = 2 , A : float = 0.02 , A : float = 1E-12 , A : bool = False , **A : Any , ): '''simple docstring''' super().__init__(pad_token_id=A , **A ) a : Optional[int] = attention_window a : Union[str, Any] = sep_token_id a : List[str] = bos_token_id a : Dict = eos_token_id a : str = vocab_size a : Union[str, Any] = hidden_size a : Any = num_hidden_layers a : Optional[Any] = num_attention_heads a : Tuple = hidden_act a : Union[str, Any] = intermediate_size a : Union[str, Any] = hidden_dropout_prob a : int = attention_probs_dropout_prob a : Union[str, Any] = max_position_embeddings a : Optional[Any] = type_vocab_size a : Optional[int] = initializer_range a : str = layer_norm_eps a : Tuple = onnx_export class snake_case ( UpperCAmelCase ): def __init__( self : Optional[int] , A : "PretrainedConfig" , A : str = "default" , A : "List[PatchingSpec]" = None ): '''simple docstring''' super().__init__(A , A , A ) a : List[Any] = True @property def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' if self.task == "multiple-choice": a : List[Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: a : List[Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('global_attention_mask', dynamic_axis), ] ) @property def lowerCamelCase__ ( self : str ): '''simple docstring''' a : Optional[Any] = super().outputs if self.task == "default": a : Tuple = {0: 'batch'} return outputs @property def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' return 1E-4 @property def lowerCamelCase__ ( self : Union[str, Any] ): '''simple docstring''' return max(super().default_onnx_opset , 1_4 ) def lowerCamelCase__ ( self : Optional[Any] , A : "PreTrainedTokenizerBase" , A : int = -1 , A : int = -1 , A : bool = False , A : Optional[TensorType] = None , ): '''simple docstring''' a : Dict = super().generate_dummy_inputs( preprocessor=A , batch_size=A , seq_length=A , is_pair=A , framework=A ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly a : int = torch.zeros_like(inputs['input_ids'] ) # make every second token global a : Optional[Any] = 1 return inputs
362
"""simple docstring""" _UpperCamelCase : List[Any] = 8.31_44_62 # Unit - J mol-1 K-1 def snake_case (A_ :float , A_ :float , A_ :float ): '''simple docstring''' if moles < 0 or kelvin < 0 or volume < 0: raise ValueError('Invalid inputs. Enter positive value.' ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / volume def snake_case (A_ :float , A_ :float , A_ :float ): '''simple docstring''' if moles < 0 or kelvin < 0 or pressure < 0: raise ValueError('Invalid inputs. Enter positive value.' ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / pressure if __name__ == "__main__": from doctest import testmod testmod()
186
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _a : Optional[int]= {"configuration_vit_msn": ["VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTMSNConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : str= [ "VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTMSNModel", "ViTMSNForImageClassification", "ViTMSNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys _a : Any= _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
172
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase = version.parse(importlib_metadata.version("""nltk""")) if NLTK_VERSION >= version.Version("""3.6.4"""): from nltk import word_tokenize _UpperCAmelCase = """\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } """ _UpperCAmelCase = """\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. """ _UpperCAmelCase = """ Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: 'meteor': meteor score. Examples: >>> meteor = datasets.load_metric('meteor') >>> predictions = [\"It is a guide to action which ensures that the military always obeys the commands of the party\"] >>> references = [\"It is a guide to action that ensures that the military will forever heed Party commands\"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results[\"meteor\"], 4)) 0.6944 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def lowerCAmelCase_ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'] , reference_urls=[ 'https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score', 'https://en.wikipedia.org/wiki/METEOR', ] , ) def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" import nltk nltk.download('wordnet' ) if NLTK_VERSION >= version.Version('3.6.5' ): nltk.download('punkt' ) if NLTK_VERSION >= version.Version('3.6.6' ): nltk.download('omw-1.4' ) def lowerCAmelCase_ ( self , lowercase , lowercase , lowercase=0.9 , lowercase=3 , lowercase=0.5 ): """simple docstring""" if NLTK_VERSION >= version.Version('3.6.5' ): A_ : List[Any] = [ meteor_score.single_meteor_score( word_tokenize(lowercase ) , word_tokenize(lowercase ) , alpha=lowercase , beta=lowercase , gamma=lowercase ) for ref, pred in zip(lowercase , lowercase ) ] else: A_ : Optional[Any] = [ meteor_score.single_meteor_score(lowercase , lowercase , alpha=lowercase , beta=lowercase , gamma=lowercase ) for ref, pred in zip(lowercase , lowercase ) ] return {"meteor": np.mean(lowercase )}
140
0
class _lowerCamelCase: """simple docstring""" def __init__( self, lowerCamelCase) -> str: """simple docstring""" _lowercase : str = arr.split(',') def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : List[str] = [int(self.array[0])] * len(self.array) _lowercase : str = [int(self.array[0])] * len(self.array) for i in range(1, len(self.array)): _lowercase : Optional[Any] = max( int(self.array[i]) + sum_value[i - 1], int(self.array[i])) _lowercase : Tuple = max(sum_value[i], rear[i - 1]) return rear[len(self.array) - 1] if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[Any] = input("please input some numbers:") SCREAMING_SNAKE_CASE : Tuple = SubArray(whole_array) SCREAMING_SNAKE_CASE : Optional[int] = array.solve_sub_array() print(("the results is:", re))
353
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Union[str, Any] = len(lowerCamelCase_ ) // 2 # choose the middle 3 elements _lowercase : Any = lst[m - 1 : m + 2] # if middle element is peak if three[1] > three[0] and three[1] > three[2]: return three[1] # if increasing, recurse on right elif three[0] < three[2]: if len(lst[:m] ) == 2: m -= 1 return peak(lst[m:] ) # decreasing else: if len(lst[:m] ) == 2: m += 1 return peak(lst[:m] ) if __name__ == "__main__": import doctest doctest.testmod()
84
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class lowerCamelCase__ ( __lowercase): '''simple docstring''' _A = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
232
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = '▁' lowercase : Tuple = {'vocab_file': 'spiece.model'} lowercase : Dict = { 'vocab_file': { 'google/reformer-crime-and-punishment': ( 'https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model' ) } } lowercase : Any = { 'google/reformer-crime-and-punishment': 524288, } class lowerCamelCase__ ( __lowercase): '''simple docstring''' _A = VOCAB_FILES_NAMES _A = PRETRAINED_VOCAB_FILES_MAP _A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A = ['input_ids', 'attention_mask'] def __init__( self :int , a :List[Any] , a :Tuple="</s>" , a :str="<unk>" , a :Dict=[] , a :Optional[Dict[str, Any]] = None , **a :Union[str, Any] , ) -> None: __UpperCamelCase : str = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=a , unk_token=a , additional_special_tokens=a , sp_model_kwargs=self.sp_model_kwargs , **a , ) __UpperCamelCase : Optional[Any] = vocab_file __UpperCamelCase : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(a ) @property def _lowerCamelCase ( self :Optional[Any] ) -> Any: return self.sp_model.get_piece_size() def _lowerCamelCase ( self :Optional[int] ) -> Dict[str, int]: __UpperCamelCase : str = {self.convert_ids_to_tokens(a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self :str ) -> List[str]: __UpperCamelCase : Union[str, Any] = self.__dict__.copy() __UpperCamelCase : Optional[Any] = None return state def __setstate__( self :int , a :List[str] ) -> int: __UpperCamelCase : Union[str, Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __UpperCamelCase : int = {} __UpperCamelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowerCamelCase ( self :List[Any] , a :str ) -> List[str]: return self.sp_model.encode(a , out_type=a ) def _lowerCamelCase ( self :Optional[int] , a :Optional[Any] ) -> str: return self.sp_model.piece_to_id(a ) def _lowerCamelCase ( self :Dict , a :Union[str, Any] ) -> Optional[int]: if index < self.sp_model.get_piece_size(): __UpperCamelCase : Optional[int] = self.sp_model.IdToPiece(a ) return token def _lowerCamelCase ( self :Dict , a :List[Any] ) -> Dict: __UpperCamelCase : Optional[int] = [] __UpperCamelCase : str = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(a ) + token __UpperCamelCase : List[Any] = [] else: current_sub_tokens.append(a ) out_string += self.sp_model.decode(a ) return out_string.strip() def _lowerCamelCase ( self :Optional[Any] , a :str , a :Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __UpperCamelCase : List[Any] = os.path.join( a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , a ) elif not os.path.isfile(self.vocab_file ): with open(a , "wb" ) as fi: __UpperCamelCase : int = self.sp_model.serialized_model_proto() fi.write(a ) return (out_vocab_file,)
232
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig _UpperCAmelCase : str = { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/config.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/config.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/config.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/config.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/config.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/config.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/config.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/config.json""", } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """albert""" def __init__( self : int , UpperCAmelCase : int=30000 , UpperCAmelCase : Dict=128 , UpperCAmelCase : List[Any]=4096 , UpperCAmelCase : Dict=12 , UpperCAmelCase : int=1 , UpperCAmelCase : Any=64 , UpperCAmelCase : Dict=16384 , UpperCAmelCase : List[str]=1 , UpperCAmelCase : int="gelu_new" , UpperCAmelCase : Tuple=0 , UpperCAmelCase : Optional[int]=0 , UpperCAmelCase : Optional[Any]=512 , UpperCAmelCase : Optional[Any]=2 , UpperCAmelCase : Any=0.0_2 , UpperCAmelCase : Any=1e-12 , UpperCAmelCase : List[str]=0.1 , UpperCAmelCase : int="absolute" , UpperCAmelCase : Optional[int]=0 , UpperCAmelCase : Optional[Any]=2 , UpperCAmelCase : Optional[int]=3 , **UpperCAmelCase : Any , ) -> int: super().__init__(pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , **UpperCAmelCase ) lowerCamelCase__ : int = vocab_size lowerCamelCase__ : Optional[int] = embedding_size lowerCamelCase__ : List[Any] = hidden_size lowerCamelCase__ : Any = num_hidden_layers lowerCamelCase__ : int = num_hidden_groups lowerCamelCase__ : str = num_attention_heads lowerCamelCase__ : Dict = inner_group_num lowerCamelCase__ : Dict = hidden_act lowerCamelCase__ : Optional[int] = intermediate_size lowerCamelCase__ : str = hidden_dropout_prob lowerCamelCase__ : Tuple = attention_probs_dropout_prob lowerCamelCase__ : str = max_position_embeddings lowerCamelCase__ : Union[str, Any] = type_vocab_size lowerCamelCase__ : Tuple = initializer_range lowerCamelCase__ : List[Any] = layer_norm_eps lowerCamelCase__ : Optional[int] = classifier_dropout_prob lowerCamelCase__ : List[str] = position_embedding_type class lowerCAmelCase ( __UpperCamelCase ): @property def A_ ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCamelCase__ : Union[str, Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowerCamelCase__ : List[str] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ] )
354
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class lowerCAmelCase ( unittest.TestCase ): def __init__( self : Dict , UpperCAmelCase : List[str] , UpperCAmelCase : Any=7 , UpperCAmelCase : int=3 , UpperCAmelCase : Optional[Any]=18 , UpperCAmelCase : str=30 , UpperCAmelCase : List[str]=400 , UpperCAmelCase : int=True , UpperCAmelCase : List[Any]=None , UpperCAmelCase : str=True , UpperCAmelCase : Optional[int]=None , UpperCAmelCase : Union[str, Any]=True , UpperCAmelCase : List[str]=[0.5, 0.5, 0.5] , UpperCAmelCase : Union[str, Any]=[0.5, 0.5, 0.5] , ) -> Tuple: lowerCamelCase__ : Union[str, Any] = size if size is not None else {'shortest_edge': 18} lowerCamelCase__ : Union[str, Any] = crop_size if crop_size is not None else {'height': 18, 'width': 18} lowerCamelCase__ : Optional[Any] = parent lowerCamelCase__ : int = batch_size lowerCamelCase__ : str = num_channels lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : List[str] = min_resolution lowerCamelCase__ : Union[str, Any] = max_resolution lowerCamelCase__ : Optional[int] = do_resize lowerCamelCase__ : int = size lowerCamelCase__ : Optional[int] = do_center_crop lowerCamelCase__ : str = crop_size lowerCamelCase__ : Optional[Any] = do_normalize lowerCamelCase__ : Tuple = image_mean lowerCamelCase__ : Union[str, Any] = image_std def A_ ( self : Any ) -> Optional[int]: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class lowerCAmelCase ( __UpperCamelCase, unittest.TestCase ): UpperCAmelCase__ = LevitImageProcessor if is_vision_available() else None def A_ ( self : Tuple ) -> Tuple: lowerCamelCase__ : str = LevitImageProcessingTester(self ) @property def A_ ( self : Tuple ) -> int: return self.image_processor_tester.prepare_image_processor_dict() def A_ ( self : Optional[int] ) -> int: lowerCamelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase , 'image_mean' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'image_std' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_normalize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_resize' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'do_center_crop' ) ) self.assertTrue(hasattr(UpperCAmelCase , 'size' ) ) def A_ ( self : List[Any] ) -> Optional[int]: lowerCamelCase__ : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18} ) self.assertEqual(image_processor.crop_size , {'height': 18, 'width': 18} ) lowerCamelCase__ : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'shortest_edge': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) def A_ ( self : str ) -> str: pass def A_ ( self : Optional[int] ) -> List[Any]: # Initialize image_processing lowerCamelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase__ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , Image.Image ) # Test not batched input lowerCamelCase__ : Optional[int] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched lowerCamelCase__ : Optional[Any] = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def A_ ( self : List[str] ) -> List[Any]: # Initialize image_processing lowerCamelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase , numpify=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , np.ndarray ) # Test not batched input lowerCamelCase__ : Tuple = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched lowerCamelCase__ : Any = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def A_ ( self : str ) -> int: # Initialize image_processing lowerCamelCase__ : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase , torchify=UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase , torch.Tensor ) # Test not batched input lowerCamelCase__ : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched lowerCamelCase__ : Any = image_processing(UpperCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
45
0
from __future__ import annotations class UpperCamelCase_ : '''simple docstring''' def __init__( self : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : str) ->List[str]: '''simple docstring''' A__ , A__ = text, pattern A__ , A__ = len(UpperCAmelCase__), len(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : str) ->int: '''simple docstring''' for i in range(self.patLen - 1 , -1 , -1): if char == self.pattern[i]: return i return -1 def SCREAMING_SNAKE_CASE ( self : int , UpperCAmelCase__ : int) ->int: '''simple docstring''' for i in range(self.patLen - 1 , -1 , -1): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def SCREAMING_SNAKE_CASE ( self : int) ->list[int]: '''simple docstring''' A__ = [] for i in range(self.textLen - self.patLen + 1): A__ = self.mismatch_in_text(UpperCAmelCase__) if mismatch_index == -1: positions.append(UpperCAmelCase__) else: A__ = self.match_in_pattern(self.text[mismatch_index]) A__ = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _lowerCamelCase : List[Any] = """ABAABA""" _lowerCamelCase : List[Any] = """AB""" _lowerCamelCase : Dict = BoyerMooreSearch(text, pattern) _lowerCamelCase : List[Any] = bms.bad_character_heuristic() if len(positions) == 0: print("""No match found""") else: print("""Pattern found in following positions: """) print(positions)
14
'''simple docstring''' import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCAmelCase_ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): '''simple docstring''' @register_to_config def __init__( self : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : float , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : str , _UpperCAmelCase : bool = False , ): """simple docstring""" super().__init__() UpperCAmelCase__ = nn.Embedding(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase__ = nn.Embedding(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase__ = False UpperCAmelCase__ = nn.Dropout(p=_UpperCAmelCase ) UpperCAmelCase__ = TaConfig( vocab_size=_UpperCAmelCase , d_model=_UpperCAmelCase , num_heads=_UpperCAmelCase , d_kv=_UpperCAmelCase , d_ff=_UpperCAmelCase , dropout_rate=_UpperCAmelCase , feed_forward_proj=_UpperCAmelCase , is_decoder=_UpperCAmelCase , is_encoder_decoder=_UpperCAmelCase , ) UpperCAmelCase__ = nn.ModuleList() for lyr_num in range(_UpperCAmelCase ): UpperCAmelCase__ = TaBlock(_UpperCAmelCase ) self.encoders.append(_UpperCAmelCase ) UpperCAmelCase__ = TaLayerNorm(_UpperCAmelCase ) UpperCAmelCase__ = nn.Dropout(p=_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Dict , _UpperCAmelCase : int , _UpperCAmelCase : str ): """simple docstring""" UpperCAmelCase__ = self.token_embedder(_UpperCAmelCase ) UpperCAmelCase__ = encoder_input_tokens.shape[1] UpperCAmelCase__ = torch.arange(_UpperCAmelCase , device=encoder_input_tokens.device ) x += self.position_encoding(_UpperCAmelCase ) UpperCAmelCase__ = self.dropout_pre(_UpperCAmelCase ) # inverted the attention mask UpperCAmelCase__ = encoder_input_tokens.size() UpperCAmelCase__ = self.get_extended_attention_mask(_UpperCAmelCase , _UpperCAmelCase ) for lyr in self.encoders: UpperCAmelCase__ = lyr(_UpperCAmelCase , _UpperCAmelCase )[0] UpperCAmelCase__ = self.layer_norm(_UpperCAmelCase ) return self.dropout_post(_UpperCAmelCase ), encoder_inputs_mask
346
0
import argparse import torch from transformers import GPTaLMHeadModel, RobertaForMaskedLM if __name__ == "__main__": snake_case_ = argparse.ArgumentParser( description=( 'Extraction some layers of the full RobertaForMaskedLM or GPT2LMHeadModel for Transfer Learned' ' Distillation' ) ) parser.add_argument('--model_type', default='roberta', choices=['roberta', 'gpt2']) parser.add_argument('--model_name', default='roberta-large', type=str) parser.add_argument('--dump_checkpoint', default='serialization_dir/tf_roberta_048131723.pth', type=str) parser.add_argument('--vocab_transform', action='store_true') snake_case_ = parser.parse_args() if args.model_type == "roberta": snake_case_ = RobertaForMaskedLM.from_pretrained(args.model_name) snake_case_ = 'roberta' elif args.model_type == "gpt2": snake_case_ = GPTaLMHeadModel.from_pretrained(args.model_name) snake_case_ = 'transformer' snake_case_ = model.state_dict() snake_case_ = {} # Embeddings # if args.model_type == "gpt2": for param_name in ["wte.weight", "wpe.weight"]: snake_case_ = state_dict[F'{prefix}.{param_name}'] else: for w in ["word_embeddings", "position_embeddings", "token_type_embeddings"]: snake_case_ = F'{prefix}.embeddings.{w}.weight' snake_case_ = state_dict[param_name] for w in ["weight", "bias"]: snake_case_ = F'{prefix}.embeddings.LayerNorm.{w}' snake_case_ = state_dict[param_name] # Transformer Blocks # snake_case_ = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: if args.model_type == "gpt2": for layer in ["ln_1", "attn.c_attn", "attn.c_proj", "ln_2", "mlp.c_fc", "mlp.c_proj"]: for w in ["weight", "bias"]: snake_case_ = state_dict[ F'{prefix}.h.{teacher_idx}.{layer}.{w}' ] snake_case_ = state_dict[F'{prefix}.h.{teacher_idx}.attn.bias'] else: for layer in [ "attention.self.query", "attention.self.key", "attention.self.value", "attention.output.dense", "attention.output.LayerNorm", "intermediate.dense", "output.dense", "output.LayerNorm", ]: for w in ["weight", "bias"]: snake_case_ = state_dict[ F'{prefix}.encoder.layer.{teacher_idx}.{layer}.{w}' ] std_idx += 1 # Language Modeling Head ###s if args.model_type == "roberta": for layer in ["lm_head.decoder.weight", "lm_head.bias"]: snake_case_ = state_dict[F'{layer}'] if args.vocab_transform: for w in ["weight", "bias"]: snake_case_ = state_dict[F'lm_head.dense.{w}'] snake_case_ = state_dict[F'lm_head.layer_norm.{w}'] elif args.model_type == "gpt2": for w in ["weight", "bias"]: snake_case_ = state_dict[F'{prefix}.ln_f.{w}'] snake_case_ = state_dict['lm_head.weight'] print(F'N layers selected for distillation: {std_idx}') print(F'Number of params transferred for distillation: {len(compressed_sd.keys())}') print(F'Save transferred checkpoint to {args.dump_checkpoint}.') torch.save(compressed_sd, args.dump_checkpoint)
369
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): A_ : Union[str, Any] = (DPMSolverSinglestepScheduler,) A_ : Union[str, Any] = (('num_inference_steps', 25),) def a (self : Dict , **a__ : Tuple ): """simple docstring""" __snake_case = { '''num_train_timesteps''': 1000, '''beta_start''': 0.0_0_0_1, '''beta_end''': 0.0_2, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''prediction_type''': '''epsilon''', '''thresholding''': False, '''sample_max_value''': 1.0, '''algorithm_type''': '''dpmsolver++''', '''solver_type''': '''midpoint''', '''lambda_min_clipped''': -float('''inf''' ), '''variance_type''': None, } config.update(**a__ ) return config def a (self : str , a__ : Any=0 , **a__ : Tuple ): """simple docstring""" __snake_case = dict(self.forward_default_kwargs ) __snake_case = kwargs.pop('''num_inference_steps''' , a__ ) __snake_case = self.dummy_sample __snake_case = 0.1 * sample __snake_case = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: __snake_case = self.get_scheduler_config(**a__ ) __snake_case = scheduler_class(**a__ ) scheduler.set_timesteps(a__ ) # copy over dummy past residuals __snake_case = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(a__ ) __snake_case = scheduler_class.from_pretrained(a__ ) new_scheduler.set_timesteps(a__ ) # copy over dummy past residuals __snake_case = dummy_past_residuals[: new_scheduler.config.solver_order] __snake_case , __snake_case = sample, sample for t in range(a__ , time_step + scheduler.config.solver_order + 1 ): __snake_case = scheduler.step(a__ , a__ , a__ , **a__ ).prev_sample __snake_case = new_scheduler.step(a__ , a__ , a__ , **a__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def a (self : Union[str, Any] ): """simple docstring""" pass def a (self : List[Any] , a__ : Dict=0 , **a__ : List[str] ): """simple docstring""" __snake_case = dict(self.forward_default_kwargs ) __snake_case = kwargs.pop('''num_inference_steps''' , a__ ) __snake_case = self.dummy_sample __snake_case = 0.1 * sample __snake_case = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a__ ) scheduler.set_timesteps(a__ ) # copy over dummy past residuals (must be after setting timesteps) __snake_case = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(a__ ) __snake_case = scheduler_class.from_pretrained(a__ ) # copy over dummy past residuals new_scheduler.set_timesteps(a__ ) # copy over dummy past residual (must be after setting timesteps) __snake_case = dummy_past_residuals[: new_scheduler.config.solver_order] __snake_case = scheduler.step(a__ , a__ , a__ , **a__ ).prev_sample __snake_case = new_scheduler.step(a__ , a__ , a__ , **a__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def a (self : int , a__ : Tuple=None , **a__ : List[str] ): """simple docstring""" if scheduler is None: __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config(**a__ ) __snake_case = scheduler_class(**a__ ) __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config(**a__ ) __snake_case = scheduler_class(**a__ ) __snake_case = 10 __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter scheduler.set_timesteps(a__ ) for i, t in enumerate(scheduler.timesteps ): __snake_case = model(a__ , a__ ) __snake_case = scheduler.step(a__ , a__ , a__ ).prev_sample return sample def a (self : str ): """simple docstring""" __snake_case = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) __snake_case = 50 __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter scheduler.set_timesteps(a__ ) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:] ): __snake_case = model(a__ , a__ ) __snake_case = scheduler.step(a__ , a__ , a__ ).prev_sample __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_mean.item() - 0.2_5_7_4 ) < 1E-3 def a (self : int ): """simple docstring""" for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=a__ ) def a (self : List[str] ): """simple docstring""" __snake_case = DPMSolverSinglestepScheduler(**self.get_scheduler_config() ) __snake_case = self.full_loop(scheduler=a__ ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_mean.item() - 0.2_7_9_1 ) < 1E-3 __snake_case = DEISMultistepScheduler.from_config(scheduler.config ) __snake_case = DPMSolverMultistepScheduler.from_config(scheduler.config ) __snake_case = UniPCMultistepScheduler.from_config(scheduler.config ) __snake_case = DPMSolverSinglestepScheduler.from_config(scheduler.config ) __snake_case = self.full_loop(scheduler=a__ ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_mean.item() - 0.2_7_9_1 ) < 1E-3 def a (self : List[str] ): """simple docstring""" self.check_over_configs(thresholding=a__ ) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=a__ , prediction_type=a__ , sample_max_value=a__ , algorithm_type='''dpmsolver++''' , solver_order=a__ , solver_type=a__ , ) def a (self : Union[str, Any] ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a__ ) def a (self : Union[str, Any] ): """simple docstring""" for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=a__ , solver_type=a__ , prediction_type=a__ , algorithm_type=a__ , ) __snake_case = self.full_loop( solver_order=a__ , solver_type=a__ , prediction_type=a__ , algorithm_type=a__ , ) assert not torch.isnan(a__ ).any(), "Samples have nan numbers" def a (self : List[str] ): """simple docstring""" self.check_over_configs(lower_order_final=a__ ) self.check_over_configs(lower_order_final=a__ ) def a (self : Optional[Any] ): """simple docstring""" self.check_over_configs(lambda_min_clipped=-float('''inf''' ) ) self.check_over_configs(lambda_min_clipped=-5.1 ) def a (self : Tuple ): """simple docstring""" self.check_over_configs(variance_type=a__ ) self.check_over_configs(variance_type='''learned_range''' ) def a (self : int ): """simple docstring""" for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=a__ , time_step=0 ) def a (self : Optional[Any] ): """simple docstring""" __snake_case = self.full_loop() __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_mean.item() - 0.2_7_9_1 ) < 1E-3 def a (self : int ): """simple docstring""" __snake_case = self.full_loop(use_karras_sigmas=a__ ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_mean.item() - 0.2_2_4_8 ) < 1E-3 def a (self : Tuple ): """simple docstring""" __snake_case = self.full_loop(prediction_type='''v_prediction''' ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_mean.item() - 0.1_4_5_3 ) < 1E-3 def a (self : List[Any] ): """simple docstring""" __snake_case = self.full_loop(prediction_type='''v_prediction''' , use_karras_sigmas=a__ ) __snake_case = torch.mean(torch.abs(a__ ) ) assert abs(result_mean.item() - 0.0_6_4_9 ) < 1E-3 def a (self : int ): """simple docstring""" __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config(thresholding=a__ , dynamic_thresholding_ratio=0 ) __snake_case = scheduler_class(**a__ ) __snake_case = 10 __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter.half() scheduler.set_timesteps(a__ ) for i, t in enumerate(scheduler.timesteps ): __snake_case = model(a__ , a__ ) __snake_case = scheduler.step(a__ , a__ , a__ ).prev_sample assert sample.dtype == torch.floataa
238
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase_ : List[str] = logging.get_logger(__name__) UpperCAmelCase_ : int = { '''salesforce/blip2-opt-2.7b''': '''https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json''', } class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : Union[str, Any] = """blip_2_vision_model""" def __init__( self : Any , __lowerCamelCase : Any=1_408 , __lowerCamelCase : Any=6_144 , __lowerCamelCase : Any=39 , __lowerCamelCase : List[str]=16 , __lowerCamelCase : List[Any]=224 , __lowerCamelCase : List[str]=14 , __lowerCamelCase : Any="gelu" , __lowerCamelCase : List[str]=0.00001 , __lowerCamelCase : Dict=0.0 , __lowerCamelCase : List[str]=1E-10 , __lowerCamelCase : int=True , **__lowerCamelCase : str , ): super().__init__(**__lowerCamelCase ) UpperCamelCase :Optional[Any] = hidden_size UpperCamelCase :List[str] = intermediate_size UpperCamelCase :Optional[Any] = num_hidden_layers UpperCamelCase :Union[str, Any] = num_attention_heads UpperCamelCase :List[str] = patch_size UpperCamelCase :str = image_size UpperCamelCase :Optional[int] = initializer_range UpperCamelCase :Optional[int] = attention_dropout UpperCamelCase :Union[str, Any] = layer_norm_eps UpperCamelCase :Optional[int] = hidden_act UpperCamelCase :List[str] = qkv_bias @classmethod def _A ( cls : Tuple , __lowerCamelCase : Union[str, os.PathLike] , **__lowerCamelCase : List[str] ): cls._set_token_in_kwargs(__lowerCamelCase ) UpperCamelCase , UpperCamelCase :List[Any] = cls.get_config_dict(__lowerCamelCase , **__lowerCamelCase ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCamelCase :Union[str, Any] = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowerCamelCase , **__lowerCamelCase ) class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : int = """blip_2_qformer""" def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=30_522 , __lowerCamelCase : Any=768 , __lowerCamelCase : Union[str, Any]=12 , __lowerCamelCase : Optional[Any]=12 , __lowerCamelCase : List[str]=3_072 , __lowerCamelCase : Union[str, Any]="gelu" , __lowerCamelCase : str=0.1 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : int=512 , __lowerCamelCase : Tuple=0.02 , __lowerCamelCase : Union[str, Any]=1E-12 , __lowerCamelCase : Any=0 , __lowerCamelCase : Tuple="absolute" , __lowerCamelCase : Optional[Any]=2 , __lowerCamelCase : Optional[Any]=1_408 , **__lowerCamelCase : str , ): super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) UpperCamelCase :Tuple = vocab_size UpperCamelCase :str = hidden_size UpperCamelCase :Dict = num_hidden_layers UpperCamelCase :Tuple = num_attention_heads UpperCamelCase :Optional[Any] = hidden_act UpperCamelCase :Any = intermediate_size UpperCamelCase :Tuple = hidden_dropout_prob UpperCamelCase :List[str] = attention_probs_dropout_prob UpperCamelCase :Tuple = max_position_embeddings UpperCamelCase :Union[str, Any] = initializer_range UpperCamelCase :Tuple = layer_norm_eps UpperCamelCase :Tuple = position_embedding_type UpperCamelCase :Optional[Any] = cross_attention_frequency UpperCamelCase :Optional[Any] = encoder_hidden_size @classmethod def _A ( cls : Optional[Any] , __lowerCamelCase : Union[str, os.PathLike] , **__lowerCamelCase : int ): cls._set_token_in_kwargs(__lowerCamelCase ) UpperCamelCase , UpperCamelCase :str = cls.get_config_dict(__lowerCamelCase , **__lowerCamelCase ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": UpperCamelCase :str = config_dict["""qformer_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__lowerCamelCase , **__lowerCamelCase ) class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : int = """blip-2""" snake_case__ : Optional[Any] = True def __init__( self : int , __lowerCamelCase : Dict=None , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[str]=None , __lowerCamelCase : Dict=32 , **__lowerCamelCase : Dict ): super().__init__(**__lowerCamelCase ) if vision_config is None: UpperCamelCase :List[str] = {} logger.info("""vision_config is None. initializing the Blip2VisionConfig with default values.""" ) if qformer_config is None: UpperCamelCase :List[Any] = {} logger.info("""qformer_config is None. Initializing the Blip2QFormerConfig with default values.""" ) if text_config is None: UpperCamelCase :int = {} logger.info("""text_config is None. Initializing the text config with default values (`OPTConfig`).""" ) UpperCamelCase :Tuple = BlipaVisionConfig(**__lowerCamelCase ) UpperCamelCase :str = BlipaQFormerConfig(**__lowerCamelCase ) UpperCamelCase :Any = text_config["""model_type"""] if """model_type""" in text_config else """opt""" UpperCamelCase :Union[str, Any] = CONFIG_MAPPING[text_model_type](**__lowerCamelCase ) UpperCamelCase :List[str] = self.text_config.tie_word_embeddings UpperCamelCase :Union[str, Any] = self.text_config.is_encoder_decoder UpperCamelCase :Tuple = num_query_tokens UpperCamelCase :Any = self.vision_config.hidden_size UpperCamelCase :int = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES UpperCamelCase :Union[str, Any] = 1.0 UpperCamelCase :Union[str, Any] = 0.02 @classmethod def _A ( cls : Tuple , __lowerCamelCase : BlipaVisionConfig , __lowerCamelCase : BlipaQFormerConfig , __lowerCamelCase : PretrainedConfig , **__lowerCamelCase : List[Any] , ): return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **__lowerCamelCase , ) def _A ( self : Optional[int] ): UpperCamelCase :Tuple = copy.deepcopy(self.__dict__ ) UpperCamelCase :Dict = self.vision_config.to_dict() UpperCamelCase :Optional[int] = self.qformer_config.to_dict() UpperCamelCase :Any = self.text_config.to_dict() UpperCamelCase :Union[str, Any] = self.__class__.model_type return output
38
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter UpperCAmelCase_ : Any = '''Create a default config file for Accelerate with only a few flags set.''' def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Optional[int]="no" , __magic_name__ : str = default_json_config_file , __magic_name__ : bool = False ) -> str: """simple docstring""" UpperCamelCase :Any = Path(__magic_name__ ) path.parent.mkdir(parents=__magic_name__ , exist_ok=__magic_name__ ) if path.exists(): print( f"""Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.""" ) return False UpperCamelCase :Dict = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( f"""`mixed_precision` should be one of 'no', 'fp16', 'bf16', or 'fp8'. Received {mixed_precision}""" ) UpperCamelCase :Optional[Any] = { """compute_environment""": """LOCAL_MACHINE""", """mixed_precision""": mixed_precision, } if torch.cuda.is_available(): UpperCamelCase :Union[str, Any] = torch.cuda.device_count() UpperCamelCase :List[Any] = num_gpus UpperCamelCase :Dict = False if num_gpus > 1: UpperCamelCase :Any = """MULTI_GPU""" else: UpperCamelCase :Any = """NO""" elif is_xpu_available() and use_xpu: UpperCamelCase :Optional[Any] = torch.xpu.device_count() UpperCamelCase :Optional[int] = num_xpus UpperCamelCase :int = False if num_xpus > 1: UpperCamelCase :Union[str, Any] = """MULTI_XPU""" else: UpperCamelCase :Union[str, Any] = """NO""" elif is_npu_available(): UpperCamelCase :List[Any] = torch.npu.device_count() UpperCamelCase :Optional[Any] = num_npus UpperCamelCase :Tuple = False if num_npus > 1: UpperCamelCase :Optional[Any] = """MULTI_NPU""" else: UpperCamelCase :List[Any] = """NO""" else: UpperCamelCase :Any = 0 UpperCamelCase :Optional[Any] = True UpperCamelCase :Optional[Any] = 1 UpperCamelCase :List[str] = """NO""" UpperCamelCase :int = ClusterConfig(**__magic_name__ ) config.to_json_file(__magic_name__ ) return path def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Dict , __magic_name__ : Tuple ) -> List[str]: """simple docstring""" UpperCamelCase :Dict = parser.add_parser("""default""" , parents=__magic_name__ , help=__magic_name__ , formatter_class=__magic_name__ ) parser.add_argument( """--config_file""" , default=__magic_name__ , help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have """ """such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed """ """with 'huggingface'.""" ) , dest="""save_location""" , ) parser.add_argument( """--mixed_precision""" , choices=["""no""", """fp16""", """bf16"""] , type=__magic_name__ , help="""Whether or not to use mixed precision training. """ """Choose between FP16 and BF16 (bfloat16) training. """ """BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later.""" , default="""no""" , ) parser.set_defaults(func=__magic_name__ ) return parser def SCREAMING_SNAKE_CASE_ ( __magic_name__ : List[Any] ) -> List[str]: """simple docstring""" UpperCamelCase :Optional[Any] = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(f"""accelerate configuration saved at {config_file}""" )
38
1
'''simple docstring''' import os import pytest from attr import dataclass __lowerCAmelCase = "us-east-1" # defaults region @dataclass class _lowerCAmelCase : '''simple docstring''' lowerCAmelCase_ = 42 lowerCAmelCase_ = "arn:aws:iam::558105141721:role/sagemaker_execution_role" lowerCAmelCase_ = { "task_name": "mnli", "per_device_train_batch_size": 16, "per_device_eval_batch_size": 16, "do_train": True, "do_eval": True, "do_predict": True, "output_dir": "/opt/ml/model", "overwrite_output_dir": True, "max_steps": 5_00, "save_steps": 55_00, } lowerCAmelCase_ = {**hyperparameters, "max_steps": 10_00} @property def lowercase (self ) -> List[Any]: if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def lowercase (self ) -> Optional[int]: return f"""{self.framework}-transfromers-test""" @property def lowercase (self ) -> List[str]: return f"""./tests/sagemaker/scripts/{self.framework}""" @property def lowercase (self ) -> List[Any]: if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope="""class""" ) def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = SageMakerTestEnvironment(framework=request.cls.framework )
361
'''simple docstring''' from __future__ import annotations import typing from collections import Counter def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): _snake_case = Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(_SCREAMING_SNAKE_CASE , max_perimeter + 1 ): _snake_case = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(_SCREAMING_SNAKE_CASE ): _snake_case = int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE = 1000 ): _snake_case = pythagorean_triple(_SCREAMING_SNAKE_CASE ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(f'''Perimeter {solution()} has maximum solutions''')
270
0
from __future__ import annotations from math import pi, sqrt def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): if inductance <= 0: raise ValueError("Inductance cannot be 0 or negative" ) elif capacitance <= 0: raise ValueError("Capacitance cannot be 0 or negative" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
19
def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = [0 for i in range(r + 1 )] # nc0 = 1 lowerCamelCase_ = 1 for i in range(1 , n + 1 ): # to compute current row from previous row. lowerCamelCase_ = min(lowerCamelCase__ , lowerCamelCase__ ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=1_0, r=5))
19
1
from __future__ import annotations class a_ : '''simple docstring''' def __init__( self : List[str] , lowercase__ : list[list[int]]): '''simple docstring''' lowerCAmelCase__ = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.') if len(_A) != 0: lowerCAmelCase__ = len(rows[0]) if cols == 0: raise error for row in rows: if len(_A) != cols: raise error for value in row: if not isinstance(_A , (int, float)): raise error lowerCAmelCase__ = rows else: lowerCAmelCase__ = [] def __snake_case ( self : Union[str, Any]): '''simple docstring''' return [[row[i] for row in self.rows] for i in range(len(self.rows[0]))] @property def __snake_case ( self : Optional[Any]): '''simple docstring''' return len(self.rows) @property def __snake_case ( self : Tuple): '''simple docstring''' return len(self.rows[0]) @property def __snake_case ( self : Tuple): '''simple docstring''' return (self.num_rows, self.num_columns) @property def __snake_case ( self : Tuple): '''simple docstring''' return self.order[0] == self.order[1] def __snake_case ( self : Tuple): '''simple docstring''' lowerCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows)] for row_num in range(self.num_rows) ] return Matrix(_A) def __snake_case ( self : Optional[int]): '''simple docstring''' if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0]) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0])) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns)) def __snake_case ( self : Dict): '''simple docstring''' return bool(self.determinant()) def __snake_case ( self : Dict , lowercase__ : int , lowercase__ : int): '''simple docstring''' lowerCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns) if other_column != column ] for other_row in range(self.num_rows) if other_row != row ] return Matrix(_A).determinant() def __snake_case ( self : str , lowercase__ : int , lowercase__ : int): '''simple docstring''' if (row + column) % 2 == 0: return self.get_minor(_A , _A) return -1 * self.get_minor(_A , _A) def __snake_case ( self : Union[str, Any]): '''simple docstring''' return Matrix( [ [self.get_minor(_A , _A) for column in range(self.num_columns)] for row in range(self.num_rows) ]) def __snake_case ( self : Any): '''simple docstring''' return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns) ] for row in range(self.minors().num_rows) ]) def __snake_case ( self : int): '''simple docstring''' lowerCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns)] for row in range(self.num_rows) ] return Matrix(_A) def __snake_case ( self : Union[str, Any]): '''simple docstring''' lowerCAmelCase__ = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse') return self.adjugate() * (1 / determinant) def __repr__( self : List[str]): '''simple docstring''' return str(self.rows) def __str__( self : Optional[int]): '''simple docstring''' if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0])) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(_A) for value in row]) + '.]' for row in self.rows ]) + "]" ) def __snake_case ( self : Optional[Any] , lowercase__ : list[int] , lowercase__ : int | None = None): '''simple docstring''' lowerCAmelCase__ = TypeError('Row must be a list containing all ints and/or floats') if not isinstance(_A , _A): raise type_error for value in row: if not isinstance(_A , (int, float)): raise type_error if len(_A) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix') if position is None: self.rows.append(_A) else: lowerCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def __snake_case ( self : List[str] , lowercase__ : list[int] , lowercase__ : int | None = None): '''simple docstring''' lowerCAmelCase__ = TypeError( 'Column must be a list containing all ints and/or floats') if not isinstance(_A , _A): raise type_error for value in column: if not isinstance(_A , (int, float)): raise type_error if len(_A) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix') if position is None: lowerCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows)] else: lowerCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows) ] def __eq__( self : Union[str, Any] , lowercase__ : object): '''simple docstring''' if not isinstance(_A , _A): return NotImplemented return self.rows == other.rows def __ne__( self : Optional[Any] , lowercase__ : object): '''simple docstring''' return not self == other def __neg__( self : List[Any]): '''simple docstring''' return self * -1 def __add__( self : List[str] , lowercase__ : Matrix): '''simple docstring''' if self.order != other.order: raise ValueError('Addition requires matrices of the same order') return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns)] for i in range(self.num_rows) ]) def __sub__( self : Tuple , lowercase__ : Matrix): '''simple docstring''' if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order') return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns)] for i in range(self.num_rows) ]) def __mul__( self : Union[str, Any] , lowercase__ : Matrix | int | float): '''simple docstring''' if isinstance(_A , (int, float)): return Matrix( [[int(element * other) for element in row] for row in self.rows]) elif isinstance(_A , _A): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second') return Matrix( [ [Matrix.dot_product(_A , _A) for column in other.columns()] for row in self.rows ]) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix') def __pow__( self : Any , lowercase__ : int): '''simple docstring''' if not isinstance(_A , _A): raise TypeError('A Matrix can only be raised to the power of an int') if not self.is_square: raise ValueError('Only square matrices can be raised to a power') if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power') lowerCAmelCase__ = self for _ in range(other - 1): result *= self return result @classmethod def __snake_case ( cls : List[str] , lowercase__ : list[int] , lowercase__ : list[int]): '''simple docstring''' return sum(row[i] * column[i] for i in range(len(_A))) if __name__ == "__main__": import doctest doctest.testmod()
357
lowerCAmelCase__ = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def __lowerCamelCase ( lowerCAmelCase__ ): # Make sure the supplied data is a bytes-like object if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase__ = F"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(lowerCAmelCase__ ) lowerCAmelCase__ = ''.join(bin(lowerCAmelCase__ )[2:].zfill(8 ) for byte in data ) lowerCAmelCase__ = len(lowerCAmelCase__ ) % 6 != 0 if padding_needed: # The padding that will be added later lowerCAmelCase__ = b'=' * ((6 - len(lowerCAmelCase__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowerCAmelCase__ ) % 6) else: lowerCAmelCase__ = b'' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowerCAmelCase__ ) , 6 ) ).encode() + padding ) def __lowerCamelCase ( lowerCAmelCase__ ): # Make sure encoded_data is either a string or a bytes-like object if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase__ = ( 'argument should be a bytes-like object or ASCII string, ' F"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(lowerCAmelCase__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): try: lowerCAmelCase__ = encoded_data.decode('utf-8' ) except UnicodeDecodeError: raise ValueError('base64 encoded data should only contain ASCII characters' ) lowerCAmelCase__ = encoded_data.count('=' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowerCAmelCase__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowerCAmelCase__ = encoded_data[:-padding] lowerCAmelCase__ = ''.join( bin(B64_CHARSET.index(lowerCAmelCase__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowerCAmelCase__ = ''.join( bin(B64_CHARSET.index(lowerCAmelCase__ ) )[2:].zfill(6 ) for char in encoded_data ) lowerCAmelCase__ = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowerCAmelCase__ ) , 8 ) ] return bytes(lowerCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
119
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: UpperCAmelCase__ : List[Any] = None UpperCAmelCase__ : Any = logging.get_logger(__name__) UpperCAmelCase__ : int = {"vocab_file": "sentencepiece.model", "tokenizer_file": "tokenizer.json"} UpperCAmelCase__ : List[Any] = { "vocab_file": { "google/rembert": "https://huggingface.co/google/rembert/resolve/main/sentencepiece.model", }, "tokenizer_file": { "google/rembert": "https://huggingface.co/google/rembert/resolve/main/tokenizer.json", }, } UpperCAmelCase__ : Dict = { "google/rembert": 2_5_6, } UpperCAmelCase__ : Tuple = "▁" class lowerCAmelCase_ (SCREAMING_SNAKE_CASE__ ): """simple docstring""" __UpperCamelCase : Tuple = VOCAB_FILES_NAMES __UpperCamelCase : Tuple = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : Tuple = RemBertTokenizer def __init__(self , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__="[CLS]" , SCREAMING_SNAKE_CASE__="[SEP]" , SCREAMING_SNAKE_CASE__="<unk>" , SCREAMING_SNAKE_CASE__="[SEP]" , SCREAMING_SNAKE_CASE__="<pad>" , SCREAMING_SNAKE_CASE__="[CLS]" , SCREAMING_SNAKE_CASE__="[MASK]" , **SCREAMING_SNAKE_CASE__ , ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = AddedToken(A__ , lstrip=A__ , rstrip=A__ ) if isinstance(A__ , A__ ) else mask_token super().__init__( A__ , tokenizer_file=A__ , do_lower_case=A__ , remove_space=A__ , keep_accents=A__ , bos_token=A__ , eos_token=A__ , unk_token=A__ , sep_token=A__ , pad_token=A__ , cls_token=A__ , mask_token=A__ , **A__ , ) SCREAMING_SNAKE_CASE__ : Optional[int] = do_lower_case SCREAMING_SNAKE_CASE__ : str = remove_space SCREAMING_SNAKE_CASE__ : str = keep_accents SCREAMING_SNAKE_CASE__ : Optional[int] = vocab_file SCREAMING_SNAKE_CASE__ : Dict = False if not self.vocab_file else True def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = [self.sep_token_id] SCREAMING_SNAKE_CASE__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = False ) -> Dict: """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( """You should not supply a second sequence if the provided sequence of """ """ids is already formatted with special tokens for the model.""" ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(A__ )) + [1] + ([0] * len(A__ )) + [1] return [1] + ([0] * len(A__ )) + [1] def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = [self.sep_token_id] SCREAMING_SNAKE_CASE__ : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ) -> int: """simple docstring""" if not os.path.isdir(A__ ): logger.error("""Vocabulary path ({}) should be a directory""".format(A__ ) ) return SCREAMING_SNAKE_CASE__ : Union[str, Any] = os.path.join( A__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A__ ): copyfile(self.vocab_file , A__ ) return (out_vocab_file,)
25
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = ('''dense.weight''', '''attention.self.query''', '''attention.self.key''', '''attention.self.value''') lowercase = ( ('''layer.''', '''layer_'''), ('''word_embeddings.weight''', '''word_embeddings'''), ('''position_embeddings.weight''', '''position_embeddings'''), ('''token_type_embeddings.weight''', '''token_type_embeddings'''), ('''.''', '''/'''), ('''LayerNorm/weight''', '''LayerNorm/gamma'''), ('''LayerNorm/bias''', '''LayerNorm/beta'''), ('''weight''', '''kernel'''), ) if not os.path.isdir(lowerCAmelCase__ ): os.makedirs(lowerCAmelCase__ ) lowercase = model.state_dict() def to_tf_var_name(lowerCAmelCase__ ): for patt, repl in iter(lowerCAmelCase__ ): lowercase = name.replace(lowerCAmelCase__ , lowerCAmelCase__ ) return f'bert/{name}' def create_tf_var(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = tf.dtypes.as_dtype(tensor.dtype ) lowercase = tf.get_variable(dtype=lowerCAmelCase__ , shape=tensor.shape , name=lowerCAmelCase__ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(lowerCAmelCase__ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase = to_tf_var_name(lowerCAmelCase__ ) lowercase = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase = torch_tensor.T lowercase = create_tf_var(tensor=lowerCAmelCase__ , name=lowerCAmelCase__ , session=lowerCAmelCase__ ) tf.keras.backend.set_value(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = session.run(lowerCAmelCase__ ) print(f'Successfully created {tf_name}: {np.allclose(lowerCAmelCase__ , lowerCAmelCase__ )}' ) lowercase = tf.train.Saver(tf.trainable_variables() ) saver.save(lowerCAmelCase__ , os.path.join(lowerCAmelCase__ , model_name.replace('''-''' , '''_''' ) + '''.ckpt''' ) ) def UpperCamelCase ( lowerCAmelCase__=None ): '''simple docstring''' lowercase = argparse.ArgumentParser() parser.add_argument('''--model_name''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''model name e.g. bert-base-uncased''' ) parser.add_argument( '''--cache_dir''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''Directory containing pytorch model''' ) parser.add_argument('''--pytorch_model_path''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''/path/to/<pytorch-model-name>.bin''' ) parser.add_argument('''--tf_cache_dir''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''Directory in which to save tensorflow model''' ) lowercase = parser.parse_args(lowerCAmelCase__ ) lowercase = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=lowerCAmelCase__ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
101
0
# Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def A ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Dict = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, nicht wahr?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] _lowerCAmelCase : Any = { '''wmt16-en-de-dist-12-1''': [28.3, 27.52], '''wmt16-en-de-dist-6-1''': [27.4, 27.11], '''wmt16-en-de-12-1''': [26.9, 25.75], } _lowerCAmelCase : int = F"{src_lang}-{tgt_lang}" _lowerCAmelCase : List[Any] = F"\n---\nlanguage:\n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt16\n- allenai\nlicense: apache-2.0\ndatasets:\n- wmt16\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}.\n\nFor more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369).\n\nAll 3 models are available:\n\n* [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1)\n* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)\n* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)\n\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"allenai/{model_name}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n\n## Training data\n\nPretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369).\n\n## Eval results\n\nHere are the BLEU scores:\n\nmodel | fairseq | transformers\n-------|---------|----------\n{model_name} | {scores[model_name][0]} | {scores[model_name][1]}\n\nThe score is slightly below the score reported in the paper, as the researchers don't use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs.\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=5\nmkdir -p $DATA_DIR\nsacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt16/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)\n\n\n### BibTeX entry and citation info\n\n```\n@misc{{kasai2020deep,\n title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},\n author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},\n year={{2020}},\n eprint={{2006.10369}},\n archivePrefix={{arXiv}},\n primaryClass={{cs.CL}}\n}}\n```\n\n" model_card_dir.mkdir(parents=lowerCamelCase_ , exist_ok=lowerCamelCase_ ) _lowerCAmelCase : Union[str, Any] = os.path.join(lowerCamelCase_ , "README.md" ) print(F"Generating {path}" ) with open(lowerCamelCase_ , "w" , encoding="utf-8" ) as f: f.write(lowerCamelCase_ ) # make sure we are under the root of the project _snake_case = Path(__file__).resolve().parent.parent.parent _snake_case = repo_dir / 'model_cards' for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: _snake_case = model_cards_dir / 'allenai' / model_name write_model_card(model_card_dir, src_lang="en", tgt_lang="de", model_name=model_name)
356
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _snake_case = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ["ConvNextFeatureExtractor"] _snake_case = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys _snake_case = _LazyModule(__name__, globals()["__file__"], _import_structure)
300
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class A_ ( unittest.TestCase ): def __init__( self : int ,SCREAMING_SNAKE_CASE__ : Dict ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=7 ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=3 ,SCREAMING_SNAKE_CASE__ : str=1_8 ,SCREAMING_SNAKE_CASE__ : Optional[int]=3_0 ,SCREAMING_SNAKE_CASE__ : List[str]=4_0_0 ,SCREAMING_SNAKE_CASE__ : Any=True ,SCREAMING_SNAKE_CASE__ : int=None ,SCREAMING_SNAKE_CASE__ : Tuple=True ,SCREAMING_SNAKE_CASE__ : str=None ,SCREAMING_SNAKE_CASE__ : Optional[Any]=True ,SCREAMING_SNAKE_CASE__ : List[str]=[0.5, 0.5, 0.5] ,SCREAMING_SNAKE_CASE__ : List[Any]=[0.5, 0.5, 0.5] ,): __lowerCamelCase : Any = size if size is not None else {'shortest_edge': 1_8} __lowerCamelCase : int = crop_size if crop_size is not None else {'height': 1_8, 'width': 1_8} __lowerCamelCase : Optional[Any] = parent __lowerCamelCase : Optional[Any] = batch_size __lowerCamelCase : List[Any] = num_channels __lowerCamelCase : List[Any] = image_size __lowerCamelCase : List[Any] = min_resolution __lowerCamelCase : Any = max_resolution __lowerCamelCase : Optional[int] = do_resize __lowerCamelCase : Optional[int] = size __lowerCamelCase : List[Any] = do_center_crop __lowerCamelCase : Tuple = crop_size __lowerCamelCase : List[str] = do_normalize __lowerCamelCase : Optional[Any] = image_mean __lowerCamelCase : int = image_std def lowerCAmelCase ( self : List[str]): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class A_ ( SCREAMING_SNAKE_CASE , unittest.TestCase ): _UpperCAmelCase : List[Any] = LevitImageProcessor if is_vision_available() else None def lowerCAmelCase ( self : int): __lowerCamelCase : Optional[int] = LevitImageProcessingTester(self) @property def lowerCAmelCase ( self : int): return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase ( self : List[str]): __lowerCamelCase : List[Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ ,'image_mean')) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ ,'image_std')) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ ,'do_normalize')) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ ,'do_resize')) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ ,'do_center_crop')) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ ,'size')) def lowerCAmelCase ( self : Optional[Any]): __lowerCamelCase : List[str] = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size ,{'shortest_edge': 1_8}) self.assertEqual(image_processor.crop_size ,{'height': 1_8, 'width': 1_8}) __lowerCamelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ,crop_size=8_4) self.assertEqual(image_processor.size ,{'shortest_edge': 4_2}) self.assertEqual(image_processor.crop_size ,{'height': 8_4, 'width': 8_4}) def lowerCAmelCase ( self : Any): pass def lowerCAmelCase ( self : List[str]): # Initialize image_processing __lowerCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __lowerCamelCase : Any = prepare_image_inputs(self.image_processor_tester ,equal_resolution=SCREAMING_SNAKE_CASE__) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,Image.Image) # Test not batched input __lowerCamelCase : Union[str, Any] = image_processing(image_inputs[0] ,return_tensors='pt').pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) # Test batched __lowerCamelCase : Optional[int] = image_processing(SCREAMING_SNAKE_CASE__ ,return_tensors='pt').pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) def lowerCAmelCase ( self : Dict): # Initialize image_processing __lowerCamelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __lowerCamelCase : List[Any] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=SCREAMING_SNAKE_CASE__ ,numpify=SCREAMING_SNAKE_CASE__) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,np.ndarray) # Test not batched input __lowerCamelCase : Optional[int] = image_processing(image_inputs[0] ,return_tensors='pt').pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) # Test batched __lowerCamelCase : Optional[int] = image_processing(SCREAMING_SNAKE_CASE__ ,return_tensors='pt').pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) def lowerCAmelCase ( self : str): # Initialize image_processing __lowerCamelCase : str = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __lowerCamelCase : str = prepare_image_inputs(self.image_processor_tester ,equal_resolution=SCREAMING_SNAKE_CASE__ ,torchify=SCREAMING_SNAKE_CASE__) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE__ ,torch.Tensor) # Test not batched input __lowerCamelCase : List[str] = image_processing(image_inputs[0] ,return_tensors='pt').pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,) # Test batched __lowerCamelCase : str = image_processing(SCREAMING_SNAKE_CASE__ ,return_tensors='pt').pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) ,)
73
import logging from transformers import PretrainedConfig a__ = logging.getLogger(__name__) a__ = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class UpperCAmelCase_ ( __lowercase ): """simple docstring""" UpperCAmelCase__ : Any = "bertabs" def __init__( self , _a=3_0_5_2_2 , _a=5_1_2 , _a=6 , _a=5_1_2 , _a=8 , _a=5_1_2 , _a=0.2 , _a=6 , _a=7_6_8 , _a=8 , _a=2_0_4_8 , _a=0.2 , **_a , ) -> Any: super().__init__(**_a ) _a : int = vocab_size _a : List[str] = max_pos _a : Tuple = enc_layers _a : Optional[Any] = enc_hidden_size _a : int = enc_heads _a : Optional[Any] = enc_ff_size _a : List[str] = enc_dropout _a : Tuple = dec_layers _a : Optional[Any] = dec_hidden_size _a : Optional[Any] = dec_heads _a : Optional[Any] = dec_ff_size _a : List[Any] = dec_dropout
235
0
"""simple docstring""" import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase : List[str] = logging.get_logger(__name__) def lowerCamelCase ( _UpperCamelCase : str , _UpperCamelCase : str ) -> List[str]: '''simple docstring''' __UpperCAmelCase : Dict = RobertaPreLayerNormConfig.from_pretrained( _UpperCamelCase , architectures=["""RobertaPreLayerNormForMaskedLM"""] ) # convert state_dict __UpperCAmelCase : Optional[Any] = torch.load(hf_hub_download(repo_id=_UpperCamelCase , filename="""pytorch_model.bin""" ) ) __UpperCAmelCase : Optional[int] = {} for tensor_key, tensor_value in original_state_dict.items(): # The transformer implementation gives the model a unique name, rather than overwiriting 'roberta' if tensor_key.startswith("""roberta.""" ): __UpperCAmelCase : Optional[int] = """roberta_prelayernorm.""" + tensor_key[len("""roberta.""" ) :] # The original implementation contains weights which are not used, remove them from the state_dict if tensor_key.endswith(""".self.LayerNorm.weight""" ) or tensor_key.endswith(""".self.LayerNorm.bias""" ): continue __UpperCAmelCase : int = tensor_value __UpperCAmelCase : Any = RobertaPreLayerNormForMaskedLM.from_pretrained( pretrained_model_name_or_path=_UpperCamelCase , config=_UpperCamelCase , state_dict=_UpperCamelCase ) model.save_pretrained(_UpperCamelCase ) # convert tokenizer __UpperCAmelCase : Union[str, Any] = AutoTokenizer.from_pretrained(_UpperCamelCase ) tokenizer.save_pretrained(_UpperCamelCase ) if __name__ == "__main__": UpperCAmelCase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint-repo', default=None, type=str, required=True, help='Path the official PyTorch dump, e.g. \'andreasmadsen/efficient_mlm_m0.40\'.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) UpperCAmelCase : str = parser.parse_args() convert_roberta_prelayernorm_checkpoint_to_pytorch(args.checkpoint_repo, args.pytorch_dump_folder_path)
320
"""simple docstring""" def lowerCamelCase ( _UpperCamelCase : str , _UpperCamelCase : Optional[int] ) -> Any: '''simple docstring''' __UpperCAmelCase : Optional[Any] = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def lowerCamelCase ( _UpperCamelCase : str , _UpperCamelCase : Optional[int] , _UpperCamelCase : List[Any] ) -> str: '''simple docstring''' __UpperCAmelCase : Dict = 0 while b > 0: if b & 1: __UpperCAmelCase : int = ((res % c) + (a % c)) % c a += a b >>= 1 return res
320
1
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __a ( __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = args.pruning_method UpperCAmelCase_ : int = args.threshold UpperCAmelCase_ : Optional[int] = args.model_name_or_path.rstrip("/" ) UpperCAmelCase_ : List[str] = args.target_model_path print(f"""Load fine-pruned model from {model_name_or_path}""" ) UpperCAmelCase_ : Tuple = torch.load(os.path.join(__lowerCamelCase, "pytorch_model.bin" ) ) UpperCAmelCase_ : Tuple = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: UpperCAmelCase_ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: UpperCAmelCase_ : Optional[Any] = tensor print(f"""Copied layer {name}""" ) elif "bias" in name: UpperCAmelCase_ : Optional[int] = tensor print(f"""Copied layer {name}""" ) else: if pruning_method == "magnitude": UpperCAmelCase_ : str = MagnitudeBinarizer.apply(inputs=__lowerCamelCase, threshold=__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue UpperCAmelCase_ : Optional[Any] = name[:-6] UpperCAmelCase_ : int = model[f"""{prefix_}mask_scores"""] UpperCAmelCase_ : List[Any] = TopKBinarizer.apply(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : int = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue UpperCAmelCase_ : Tuple = name[:-6] UpperCAmelCase_ : Optional[Any] = model[f"""{prefix_}mask_scores"""] UpperCAmelCase_ : Any = ThresholdBinarizer.apply(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : Tuple = tensor * mask print(f"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue UpperCAmelCase_ : Optional[int] = name[:-6] UpperCAmelCase_ : List[str] = model[f"""{prefix_}mask_scores"""] UpperCAmelCase_ , UpperCAmelCase_ : int = -0.1, 1.1 UpperCAmelCase_ : str = torch.sigmoid(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = s * (r - l) + l UpperCAmelCase_ : Any = s_bar.clamp(min=0.0, max=1.0 ) UpperCAmelCase_ : Optional[Any] = tensor * mask print(f"""Pruned layer {name}""" ) else: raise ValueError("Unknown pruning method" ) if target_model_path is None: UpperCAmelCase_ : int = os.path.join( os.path.dirname(__lowerCamelCase ), f"""bertarized_{os.path.basename(__lowerCamelCase )}""" ) if not os.path.isdir(__lowerCamelCase ): shutil.copytree(__lowerCamelCase, __lowerCamelCase ) print(f"""\nCreated folder {target_model_path}""" ) torch.save(__lowerCamelCase, os.path.join(__lowerCamelCase, "pytorch_model.bin" ) ) print("\nPruned model saved! See you later!" ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) _a = parser.parse_args() main(args)
61
'''simple docstring''' from __future__ import annotations from collections.abc import Sequence from typing import Literal def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> str | Literal[False]: _a : Optional[int] = list(lowerCAmelCase_ ) _a : Optional[Any] = list(lowerCAmelCase_ ) _a : Union[str, Any] = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count += 1 _a : Optional[int] = '_' if count > 1: return False else: return "".join(lowerCAmelCase_ ) def __lowerCamelCase ( lowerCAmelCase_ ) -> list[str]: _a : Optional[int] = [] while True: _a : Any = ['$'] * len(lowerCAmelCase_ ) _a : List[str] = [] for i in range(len(lowerCAmelCase_ ) ): for j in range(i + 1 , len(lowerCAmelCase_ ) ): _a : Optional[int] = compare_string(binary[i] , binary[j] ) if k is False: _a : Optional[Any] = '*' _a : Optional[Any] = '*' temp.append('X' ) for i in range(len(lowerCAmelCase_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowerCAmelCase_ ) == 0: return pi _a : Any = list(set(lowerCAmelCase_ ) ) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> list[str]: _a : int = [] for minterm in minterms: _a : Optional[int] = '' for _ in range(lowerCAmelCase_ ): _a : Union[str, Any] = str(minterm % 2 ) + string minterm //= 2 temp.append(lowerCAmelCase_ ) return temp def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> bool: _a : int = list(lowerCAmelCase_ ) _a : Union[str, Any] = list(lowerCAmelCase_ ) _a : str = 0 for i in range(len(lowerCAmelCase_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> list[str]: _a : List[Any] = [] _a : Optional[Any] = [0] * len(lowerCAmelCase_ ) for i in range(len(chart[0] ) ): _a : Union[str, Any] = 0 _a : int = -1 for j in range(len(lowerCAmelCase_ ) ): if chart[j][i] == 1: count += 1 _a : int = j if count == 1: _a : List[Any] = 1 for i in range(len(lowerCAmelCase_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowerCAmelCase_ ) ): _a : Any = 0 temp.append(prime_implicants[i] ) while True: _a : Union[str, Any] = 0 _a : List[Any] = -1 _a : str = 0 for i in range(len(lowerCAmelCase_ ) ): _a : Union[str, Any] = chart[i].count(1 ) if count_n > max_n: _a : Any = count_n _a : int = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowerCAmelCase_ ) ): _a : List[str] = 0 def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> list[list[int]]: _a : int = [[0 for x in range(len(lowerCAmelCase_ ) )] for x in range(len(lowerCAmelCase_ ) )] for i in range(len(lowerCAmelCase_ ) ): _a : str = prime_implicants[i].count('_' ) for j in range(len(lowerCAmelCase_ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowerCAmelCase_ ): _a : Optional[Any] = 1 return chart def __lowerCamelCase ( ) -> None: _a : Optional[int] = int(input('Enter the no. of variables\n' ) ) _a : List[Any] = [ float(lowerCAmelCase_ ) for x in input( 'Enter the decimal representation of Minterms \'Spaces Separated\'\n' ).split() ] _a : List[str] = decimal_to_binary(lowerCAmelCase_ , lowerCAmelCase_ ) _a : Dict = check(lowerCAmelCase_ ) print('Prime Implicants are:' ) print(lowerCAmelCase_ ) _a : List[Any] = prime_implicant_chart(lowerCAmelCase_ , lowerCAmelCase_ ) _a : int = selection(lowerCAmelCase_ , lowerCAmelCase_ ) print('Essential Prime Implicants are:' ) print(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
89
0
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline _SCREAMING_SNAKE_CASE = argparse.ArgumentParser('''Stable Diffusion script with intel optimization''', add_help=False) parser.add_argument('''--dpm''', action='''store_true''', help='''Enable DPMSolver or not''') parser.add_argument('''--steps''', default=None, type=int, help='''Num inference steps''') _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = '''cpu''' _SCREAMING_SNAKE_CASE = '''a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings''' _SCREAMING_SNAKE_CASE = '''path-to-your-trained-model''' _SCREAMING_SNAKE_CASE = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: _SCREAMING_SNAKE_CASE = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) _SCREAMING_SNAKE_CASE = pipe.to(device) # to channels last _SCREAMING_SNAKE_CASE = pipe.unet.to(memory_format=torch.channels_last) _SCREAMING_SNAKE_CASE = pipe.vae.to(memory_format=torch.channels_last) _SCREAMING_SNAKE_CASE = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: _SCREAMING_SNAKE_CASE = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex _SCREAMING_SNAKE_CASE = torch.randn(2, 4, 6_4, 6_4) _SCREAMING_SNAKE_CASE = torch.rand(1) * 9_9_9 _SCREAMING_SNAKE_CASE = torch.randn(2, 7_7, 7_6_8) _SCREAMING_SNAKE_CASE = (sample, timestep, encoder_hidden_status) try: _SCREAMING_SNAKE_CASE = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: _SCREAMING_SNAKE_CASE = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) _SCREAMING_SNAKE_CASE = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) _SCREAMING_SNAKE_CASE = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: _SCREAMING_SNAKE_CASE = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute _SCREAMING_SNAKE_CASE = 6_6_6 _SCREAMING_SNAKE_CASE = torch.Generator(device).manual_seed(seed) _SCREAMING_SNAKE_CASE = {'''generator''': generator} if args.steps is not None: _SCREAMING_SNAKE_CASE = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): _SCREAMING_SNAKE_CASE = pipe(prompt, **generate_kwargs).images[0] # save image image.save('''generated.png''')
217
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { '''Intel/dpt-large''': '''https://huggingface.co/Intel/dpt-large/resolve/main/config.json''', # See all DPT models at https://huggingface.co/models?filter=dpt } class __lowercase ( lowerCAmelCase__ ): '''simple docstring''' a : Union[str, Any] = "dpt" def __init__(self ,_lowerCamelCase=768 ,_lowerCamelCase=12 ,_lowerCamelCase=12 ,_lowerCamelCase=3072 ,_lowerCamelCase="gelu" ,_lowerCamelCase=0.0 ,_lowerCamelCase=0.0 ,_lowerCamelCase=0.0_2 ,_lowerCamelCase=1E-1_2 ,_lowerCamelCase=384 ,_lowerCamelCase=16 ,_lowerCamelCase=3 ,_lowerCamelCase=False ,_lowerCamelCase=True ,_lowerCamelCase=[2, 5, 8, 11] ,_lowerCamelCase="project" ,_lowerCamelCase=[4, 2, 1, 0.5] ,_lowerCamelCase=[96, 192, 384, 768] ,_lowerCamelCase=256 ,_lowerCamelCase=-1 ,_lowerCamelCase=False ,_lowerCamelCase=True ,_lowerCamelCase=0.4 ,_lowerCamelCase=255 ,_lowerCamelCase=0.1 ,_lowerCamelCase=[1, 1024, 24, 24] ,_lowerCamelCase=[0, 1] ,_lowerCamelCase=None ,**_lowerCamelCase ,) -> Tuple: '''simple docstring''' super().__init__(**_lowerCamelCase ) __lowercase = hidden_size __lowercase = is_hybrid if self.is_hybrid: if backbone_config is None: logger.info('''Initializing the config with a `BiT` backbone.''' ) __lowercase = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, } __lowercase = BitConfig(**_lowerCamelCase ) elif isinstance(_lowerCamelCase ,_lowerCamelCase ): logger.info('''Initializing the config with a `BiT` backbone.''' ) __lowercase = BitConfig(**_lowerCamelCase ) elif isinstance(_lowerCamelCase ,_lowerCamelCase ): __lowercase = backbone_config else: raise ValueError( f"backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}." ) __lowercase = backbone_featmap_shape __lowercase = neck_ignore_stages if readout_type != "project": raise ValueError('''Readout type must be \'project\' when using `DPT-hybrid` mode.''' ) else: __lowercase = None __lowercase = None __lowercase = [] __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = qkv_bias __lowercase = backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError('''Readout_type must be one of [\'ignore\', \'add\', \'project\']''' ) __lowercase = readout_type __lowercase = reassemble_factors __lowercase = neck_hidden_sizes __lowercase = fusion_hidden_size __lowercase = head_in_index __lowercase = use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) __lowercase = use_auxiliary_head __lowercase = auxiliary_loss_weight __lowercase = semantic_loss_ignore_index __lowercase = semantic_classifier_dropout def _UpperCAmelCase (self ) -> Union[str, Any]: '''simple docstring''' __lowercase = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __lowercase = self.backbone_config.to_dict() __lowercase = self.__class__.model_type return output
217
1
'''simple docstring''' def a_ ( __snake_case : int = 400_0000 ) -> int: """simple docstring""" lowerCamelCase_ =[] lowerCamelCase_, lowerCamelCase_ =0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__snake_case ) lowerCamelCase_, lowerCamelCase_ =b, a + b return sum(__snake_case ) if __name__ == "__main__": print(F"""{solution() = }""")
75
def a__ ( __UpperCamelCase ): if n == 1 or not isinstance(__UpperCamelCase , __UpperCamelCase ): return 0 elif n == 2: return 1 else: SCREAMING_SNAKE_CASE_ = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = 2 while digits < n: index += 1 SCREAMING_SNAKE_CASE_ = len(str(fibonacci(__UpperCamelCase ) ) ) return index def a__ ( __UpperCamelCase = 1_0_0_0 ): return fibonacci_digits_index(__UpperCamelCase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
118
0
'''simple docstring''' from math import factorial def lowerCamelCase (_SCREAMING_SNAKE_CASE : int = 100 ): return sum(map(_a , str(factorial(_a ) ) ) ) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
361
'''simple docstring''' import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class __UpperCamelCase ( lowerCAmelCase_ ): def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=32 , __a=5 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=512 , __a=16 , __a=2 , __a=0.02 , __a=False , __a=True , __a="None" , __a=3 , __a=4 , __a=None , ): '''simple docstring''' __a : int = parent __a : Union[str, Any] = batch_size __a : Optional[int] = seq_length __a : List[str] = is_training __a : Any = use_input_mask __a : Optional[int] = use_token_type_ids __a : Any = use_labels __a : List[str] = vocab_size __a : str = hidden_size __a : List[str] = num_hidden_layers __a : str = num_attention_heads __a : Optional[int] = intermediate_size __a : Tuple = hidden_act __a : Union[str, Any] = hidden_dropout_prob __a : Dict = attention_probs_dropout_prob __a : Optional[int] = max_position_embeddings __a : Dict = type_vocab_size __a : Any = type_sequence_label_size __a : Dict = initializer_range __a : Optional[Any] = num_labels __a : Optional[Any] = num_choices __a : Union[str, Any] = relative_attention __a : List[str] = position_biased_input __a : List[Any] = pos_att_type __a : Tuple = scope def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a : List[Any] = None if self.use_input_mask: __a : Any = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __a : Any = None if self.use_token_type_ids: __a : int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __a : Optional[int] = None __a : int = None __a : Dict = None if self.use_labels: __a : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __a : List[str] = ids_tensor([self.batch_size] , self.num_choices ) __a : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self ): '''simple docstring''' return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def __UpperCAmelCase ( self , __a ): '''simple docstring''' self.parent.assertListEqual(list(result.loss.size() ) , [] ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : Dict = DebertaVaModel(config=__a ) model.to(__a ) model.eval() __a : Optional[int] = model(__a , attention_mask=__a , token_type_ids=__a )[0] __a : str = model(__a , token_type_ids=__a )[0] __a : Optional[int] = model(__a )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : int = DebertaVaForMaskedLM(config=__a ) model.to(__a ) model.eval() __a : List[Any] = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : Optional[Any] = self.num_labels __a : List[Any] = DebertaVaForSequenceClassification(__a ) model.to(__a ) model.eval() __a : Any = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(__a ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : Any = self.num_labels __a : Dict = DebertaVaForTokenClassification(config=__a ) model.to(__a ) model.eval() __a : str = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : List[str] = DebertaVaForQuestionAnswering(config=__a ) model.to(__a ) model.eval() __a : str = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCAmelCase ( self , __a , __a , __a , __a , __a , __a , __a ): '''simple docstring''' __a : Optional[int] = DebertaVaForMultipleChoice(config=__a ) model.to(__a ) model.eval() __a : Any = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : Optional[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : Optional[int] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a : int = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) : Dict = config_and_inputs __a : Optional[int] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class __UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): A_ = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) A_ = ( { "feature-extraction": DebertaVaModel, "fill-mask": DebertaVaForMaskedLM, "question-answering": DebertaVaForQuestionAnswering, "text-classification": DebertaVaForSequenceClassification, "token-classification": DebertaVaForTokenClassification, "zero-shot": DebertaVaForSequenceClassification, } if is_torch_available() else {} ) A_ = True A_ = False A_ = False A_ = False A_ = False def __UpperCAmelCase ( self ): '''simple docstring''' __a : Union[str, Any] = DebertaVaModelTester(self ) __a : List[str] = ConfigTester(self , config_class=__a , hidden_size=37 ) def __UpperCAmelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*__a ) @slow def __UpperCAmelCase ( self ): '''simple docstring''' for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : str = DebertaVaModel.from_pretrained(__a ) self.assertIsNotNone(__a ) @require_torch @require_sentencepiece @require_tokenizers class __UpperCamelCase ( unittest.TestCase ): @unittest.skip(reason='Model not available yet' ) def __UpperCAmelCase ( self ): '''simple docstring''' pass @slow def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[int] = DebertaVaModel.from_pretrained('microsoft/deberta-v2-xlarge' ) __a : Optional[Any] = torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) __a : str = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __a : int = model(__a , attention_mask=__a )[0] # compare the actual values for a slice. __a : str = torch.tensor( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __a , atol=1E-4 ) , f"""{output[:, 1:4, 1:4]}""" )
294
0
# Imports import numpy as np class _a : """simple docstring""" def __init__( self: str , __lowerCamelCase: Optional[int]=None , __lowerCamelCase: Optional[int]=None , __lowerCamelCase: Union[str, Any]=None , __lowerCamelCase: List[str]=None , __lowerCamelCase: Dict=None ): '''simple docstring''' self.set_matricies(red=__lowerCamelCase , green=__lowerCamelCase , blue=__lowerCamelCase , red_edge=__lowerCamelCase , nir=__lowerCamelCase ) def UpperCAmelCase_ ( self: Optional[Any] , __lowerCamelCase: Optional[Any]=None , __lowerCamelCase: Optional[int]=None , __lowerCamelCase: Union[str, Any]=None , __lowerCamelCase: str=None , __lowerCamelCase: int=None ): '''simple docstring''' if red is not None: UpperCamelCase__: str = red if green is not None: UpperCamelCase__: Optional[int] = green if blue is not None: UpperCamelCase__: Union[str, Any] = blue if red_edge is not None: UpperCamelCase__: List[str] = red_edge if nir is not None: UpperCamelCase__: Dict = nir return True def UpperCAmelCase_ ( self: List[Any] , __lowerCamelCase: List[str]="" , __lowerCamelCase: Any=None , __lowerCamelCase: Union[str, Any]=None , __lowerCamelCase: Any=None , __lowerCamelCase: Optional[Any]=None , __lowerCamelCase: str=None ): '''simple docstring''' self.set_matricies(red=__lowerCamelCase , green=__lowerCamelCase , blue=__lowerCamelCase , red_edge=__lowerCamelCase , nir=__lowerCamelCase ) UpperCamelCase__: List[Any] = { "ARVI2": self.arvaa, "CCCI": self.ccci, "CVI": self.cvi, "GLI": self.gli, "NDVI": self.ndvi, "BNDVI": self.bndvi, "redEdgeNDVI": self.red_edge_ndvi, "GNDVI": self.gndvi, "GBNDVI": self.gbndvi, "GRNDVI": self.grndvi, "RBNDVI": self.rbndvi, "PNDVI": self.pndvi, "ATSAVI": self.atsavi, "BWDRVI": self.bwdrvi, "CIgreen": self.ci_green, "CIrededge": self.ci_rededge, "CI": self.ci, "CTVI": self.ctvi, "GDVI": self.gdvi, "EVI": self.evi, "GEMI": self.gemi, "GOSAVI": self.gosavi, "GSAVI": self.gsavi, "Hue": self.hue, "IVI": self.ivi, "IPVI": self.ipvi, "I": self.i, "RVI": self.rvi, "MRVI": self.mrvi, "MSAVI": self.m_savi, "NormG": self.norm_g, "NormNIR": self.norm_nir, "NormR": self.norm_r, "NGRDI": self.ngrdi, "RI": self.ri, "S": self.s, "IF": self._if, "DVI": self.dvi, "TVI": self.tvi, "NDRE": self.ndre, } try: return funcs[index]() except KeyError: print("Index not in the list!" ) return False def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' return self.nir * (self.red / (self.green**2)) def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' return (self.nir - self.red) / (self.nir + self.red) def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return (self.nir - self.blue) / (self.nir + self.blue) def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' return (self.redEdge - self.red) / (self.redEdge + self.red) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' return (self.nir - self.green) / (self.nir + self.green) def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def UpperCAmelCase_ ( self: List[Any] , __lowerCamelCase: Any=0.08 , __lowerCamelCase: List[str]=1.22 , __lowerCamelCase: Optional[Any]=0.03 ): '''simple docstring''' return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def UpperCAmelCase_ ( self: Union[str, Any] ): '''simple docstring''' return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' return (self.nir / self.green) - 1 def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' return (self.nir / self.redEdge) - 1 def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' return (self.red - self.blue) / self.red def UpperCAmelCase_ ( self: Any ): '''simple docstring''' UpperCamelCase__: Optional[Any] = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return self.nir - self.green def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def UpperCAmelCase_ ( self: Union[str, Any] ): '''simple docstring''' UpperCamelCase__: int = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.125) / (1 - self.red) def UpperCAmelCase_ ( self: Dict , __lowerCamelCase: str=0.16 ): '''simple docstring''' return (self.nir - self.green) / (self.nir + self.green + y) def UpperCAmelCase_ ( self: Optional[int] , __lowerCamelCase: int=0.5 ): '''simple docstring''' return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def UpperCAmelCase_ ( self: Union[str, Any] , __lowerCamelCase: Tuple=None , __lowerCamelCase: Union[str, Any]=None ): '''simple docstring''' return (self.nir - b) / (a * self.red) def UpperCAmelCase_ ( self: int ): '''simple docstring''' return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' return (self.red + self.green + self.blue) / 30.5 def UpperCAmelCase_ ( self: str ): '''simple docstring''' return self.nir / self.red def UpperCAmelCase_ ( self: int ): '''simple docstring''' return (self.rvi() - 1) / (self.rvi() + 1) def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' return self.green / (self.nir + self.red + self.green) def UpperCAmelCase_ ( self: str ): '''simple docstring''' return self.nir / (self.nir + self.red + self.green) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' return self.red / (self.nir + self.red + self.green) def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return (self.green - self.red) / (self.green + self.red) def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' return (self.red - self.green) / (self.red + self.green) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' UpperCamelCase__: int = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) UpperCamelCase__: Dict = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def UpperCAmelCase_ ( self: str ): '''simple docstring''' return self.nir / self.red def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' return (self.ndvi() + 0.5) ** (1 / 2) def UpperCAmelCase_ ( self: int ): '''simple docstring''' return (self.nir - self.redEdge) / (self.nir + self.redEdge)
149
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _a ( UpperCamelCase__): """simple docstring""" def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' UpperCamelCase__: str = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__lowerCamelCase , "embed_dim" ) ) self.parent.assertTrue(hasattr(__lowerCamelCase , "num_heads" ) ) class _a : """simple docstring""" def __init__( self: Tuple , __lowerCamelCase: List[str] , __lowerCamelCase: str=13 , __lowerCamelCase: Tuple=64 , __lowerCamelCase: List[Any]=3 , __lowerCamelCase: List[Any]=[16, 48, 96] , __lowerCamelCase: Union[str, Any]=[1, 3, 6] , __lowerCamelCase: Tuple=[1, 2, 10] , __lowerCamelCase: int=[7, 3, 3] , __lowerCamelCase: Dict=[4, 2, 2] , __lowerCamelCase: int=[2, 1, 1] , __lowerCamelCase: Dict=[2, 2, 2] , __lowerCamelCase: List[str]=[False, False, True] , __lowerCamelCase: str=[0.0, 0.0, 0.0] , __lowerCamelCase: Union[str, Any]=0.02 , __lowerCamelCase: str=1e-12 , __lowerCamelCase: Optional[Any]=True , __lowerCamelCase: Tuple=True , __lowerCamelCase: Union[str, Any]=2 , ): '''simple docstring''' UpperCamelCase__: Dict = parent UpperCamelCase__: Union[str, Any] = batch_size UpperCamelCase__: int = image_size UpperCamelCase__: Dict = patch_sizes UpperCamelCase__: Any = patch_stride UpperCamelCase__: Optional[int] = patch_padding UpperCamelCase__: Any = is_training UpperCamelCase__: Dict = use_labels UpperCamelCase__: List[str] = num_labels UpperCamelCase__: Tuple = num_channels UpperCamelCase__: int = embed_dim UpperCamelCase__: int = num_heads UpperCamelCase__: Dict = stride_kv UpperCamelCase__: Optional[int] = depth UpperCamelCase__: int = cls_token UpperCamelCase__: Optional[Any] = attention_drop_rate UpperCamelCase__: Tuple = initializer_range UpperCamelCase__: Dict = layer_norm_eps def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' UpperCamelCase__: Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase__: Any = None if self.use_labels: # create a random int32 tensor of given shape UpperCamelCase__: Tuple = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase__: List[Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def UpperCAmelCase_ ( self: int , __lowerCamelCase: Tuple , __lowerCamelCase: List[str] , __lowerCamelCase: Optional[Any] ): '''simple docstring''' UpperCamelCase__: str = TFCvtModel(config=__lowerCamelCase ) UpperCamelCase__: str = model(__lowerCamelCase , training=__lowerCamelCase ) UpperCamelCase__: Optional[Any] = (self.image_size, self.image_size) UpperCamelCase__ , UpperCamelCase__: Optional[Any] = image_size[0], image_size[1] for i in range(len(self.depth ) ): UpperCamelCase__: Optional[Any] = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) UpperCamelCase__: List[Any] = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def UpperCAmelCase_ ( self: Dict , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Dict ): '''simple docstring''' UpperCamelCase__: int = self.num_labels UpperCamelCase__: Tuple = TFCvtForImageClassification(__lowerCamelCase ) UpperCamelCase__: Tuple = model(__lowerCamelCase , labels=__lowerCamelCase , training=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self: Union[str, Any] ): '''simple docstring''' UpperCamelCase__: Union[str, Any] = self.prepare_config_and_inputs() UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__: Dict = config_and_inputs UpperCamelCase__: int = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class _a ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase): """simple docstring""" UpperCamelCase__ = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () UpperCamelCase__ = ( {"""feature-extraction""": TFCvtModel, """image-classification""": TFCvtForImageClassification} if is_tf_available() else {} ) UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def UpperCAmelCase_ ( self: int ): '''simple docstring''' UpperCamelCase__: Optional[Any] = TFCvtModelTester(self ) UpperCamelCase__: int = TFCvtConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason="Cvt does not output attentions" ) def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' pass @unittest.skip(reason="Cvt does not use inputs_embeds" ) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' pass @unittest.skip(reason="Cvt does not support input and output embeddings" ) def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("GPU" ) ) == 0 , reason="TF does not support backprop for grouped convolutions on CPU." , ) def UpperCAmelCase_ ( self: int ): '''simple docstring''' super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("GPU" ) ) == 0 , reason="TF does not support backprop for grouped convolutions on CPU." , ) @slow def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' super().test_keras_fit() @unittest.skip(reason="Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8" ) def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' UpperCamelCase__: int = tf.keras.mixed_precision.Policy("mixed_float16" ) tf.keras.mixed_precision.set_global_policy(__lowerCamelCase ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy("float32" ) def UpperCAmelCase_ ( self: List[str] ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__: List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__: int = model_class(__lowerCamelCase ) UpperCamelCase__: Dict = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase__: List[str] = [*signature.parameters.keys()] UpperCamelCase__: Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' def check_hidden_states_output(__lowerCamelCase: Union[str, Any] , __lowerCamelCase: Optional[int] , __lowerCamelCase: str ): UpperCamelCase__: Tuple = model_class(__lowerCamelCase ) UpperCamelCase__: Any = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) UpperCamelCase__: Optional[Any] = outputs.hidden_states UpperCamelCase__: str = len(self.model_tester.depth ) self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) UpperCamelCase__ , UpperCamelCase__: List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__: int = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase__: Union[str, Any] = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' UpperCamelCase__: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' UpperCamelCase__: Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase__: List[Any] = TFCvtModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def lowerCAmelCase_ ( ): UpperCamelCase__: Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") return image @require_tf @require_vision class _a ( unittest.TestCase): """simple docstring""" @cached_property def UpperCAmelCase_ ( self: Union[str, Any] ): '''simple docstring''' return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' UpperCamelCase__: Union[str, Any] = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCamelCase__: str = self.default_image_processor UpperCamelCase__: List[str] = prepare_img() UpperCamelCase__: Tuple = image_processor(images=__lowerCamelCase , return_tensors="tf" ) # forward pass UpperCamelCase__: List[str] = model(**__lowerCamelCase ) # verify the logits UpperCamelCase__: Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) UpperCamelCase__: Optional[Any] = tf.constant([0.9_285, 0.9_015, -0.3_150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , __lowerCamelCase , atol=1e-4 ) )
149
1
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed SCREAMING_SNAKE_CASE__ : Dict = { "distilbert": (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), "roberta": (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), "bert": (BertConfig, BertForMaskedLM, BertTokenizer), "gpt2": (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def __magic_name__ ( __lowerCAmelCase : Optional[int] ) -> List[Any]: assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def __magic_name__ ( __lowerCAmelCase : Dict , __lowerCAmelCase : Optional[int] ) -> Optional[Any]: if args.student_type == "roberta": __lowerCamelCase = False elif args.student_type == "gpt2": __lowerCamelCase = False def __magic_name__ ( __lowerCAmelCase : Any , __lowerCAmelCase : str ) -> List[Any]: if args.student_type == "roberta": __lowerCamelCase = False def __magic_name__ ( ) -> int: __lowerCamelCase = argparse.ArgumentParser(description='''Training''' ) parser.add_argument('''--force''' , action='''store_true''' , help='''Overwrite dump_path if it already exists.''' ) parser.add_argument( '''--dump_path''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''The output directory (log, checkpoints, parameters, etc.)''' ) parser.add_argument( '''--data_file''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''The binarized file (tokenized + tokens_to_ids) and grouped by sequence.''' , ) parser.add_argument( '''--student_type''' , type=__lowerCAmelCase , choices=['''distilbert''', '''roberta''', '''gpt2'''] , required=__lowerCAmelCase , help='''The student type (DistilBERT, RoBERTa).''' , ) parser.add_argument('''--student_config''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to the student configuration.''' ) parser.add_argument( '''--student_pretrained_weights''' , default=__lowerCAmelCase , type=__lowerCAmelCase , help='''Load student initialization checkpoint.''' ) parser.add_argument( '''--teacher_type''' , choices=['''bert''', '''roberta''', '''gpt2'''] , required=__lowerCAmelCase , help='''Teacher type (BERT, RoBERTa).''' ) parser.add_argument('''--teacher_name''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''The teacher model.''' ) parser.add_argument('''--temperature''' , default=2.0 , type=__lowerCAmelCase , help='''Temperature for the softmax temperature.''' ) parser.add_argument( '''--alpha_ce''' , default=0.5 , type=__lowerCAmelCase , help='''Linear weight for the distillation loss. Must be >=0.''' ) parser.add_argument( '''--alpha_mlm''' , default=0.0 , type=__lowerCAmelCase , help='''Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.''' , ) parser.add_argument('''--alpha_clm''' , default=0.5 , type=__lowerCAmelCase , help='''Linear weight for the CLM loss. Must be >=0.''' ) parser.add_argument('''--alpha_mse''' , default=0.0 , type=__lowerCAmelCase , help='''Linear weight of the MSE loss. Must be >=0.''' ) parser.add_argument( '''--alpha_cos''' , default=0.0 , type=__lowerCAmelCase , help='''Linear weight of the cosine embedding loss. Must be >=0.''' ) parser.add_argument( '''--mlm''' , action='''store_true''' , help='''The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.''' ) parser.add_argument( '''--mlm_mask_prop''' , default=0.15 , type=__lowerCAmelCase , help='''Proportion of tokens for which we need to make a prediction.''' , ) parser.add_argument('''--word_mask''' , default=0.8 , type=__lowerCAmelCase , help='''Proportion of tokens to mask out.''' ) parser.add_argument('''--word_keep''' , default=0.1 , type=__lowerCAmelCase , help='''Proportion of tokens to keep.''' ) parser.add_argument('''--word_rand''' , default=0.1 , type=__lowerCAmelCase , help='''Proportion of tokens to randomly replace.''' ) parser.add_argument( '''--mlm_smoothing''' , default=0.7 , type=__lowerCAmelCase , help='''Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).''' , ) parser.add_argument('''--token_counts''' , type=__lowerCAmelCase , help='''The token counts in the data_file for MLM.''' ) parser.add_argument( '''--restrict_ce_to_mask''' , action='''store_true''' , help='''If true, compute the distillation loss only the [MLM] prediction distribution.''' , ) parser.add_argument( '''--freeze_pos_embs''' , action='''store_true''' , help='''Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.''' , ) parser.add_argument( '''--freeze_token_type_embds''' , action='''store_true''' , help='''Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.''' , ) parser.add_argument('''--n_epoch''' , type=__lowerCAmelCase , default=3 , help='''Number of pass on the whole dataset.''' ) parser.add_argument('''--batch_size''' , type=__lowerCAmelCase , default=5 , help='''Batch size (for each process).''' ) parser.add_argument( '''--group_by_size''' , action='''store_false''' , help='''If true, group sequences that have similar length into the same batch. Default is true.''' , ) parser.add_argument( '''--gradient_accumulation_steps''' , type=__lowerCAmelCase , default=50 , help='''Gradient accumulation for larger training batches.''' , ) parser.add_argument('''--warmup_prop''' , default=0.05 , type=__lowerCAmelCase , help='''Linear warmup proportion.''' ) parser.add_argument('''--weight_decay''' , default=0.0 , type=__lowerCAmelCase , help='''Weight decay if we apply some.''' ) parser.add_argument('''--learning_rate''' , default=5E-4 , type=__lowerCAmelCase , help='''The initial learning rate for Adam.''' ) parser.add_argument('''--adam_epsilon''' , default=1E-6 , type=__lowerCAmelCase , help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' , default=5.0 , type=__lowerCAmelCase , help='''Max gradient norm.''' ) parser.add_argument('''--initializer_range''' , default=0.02 , type=__lowerCAmelCase , help='''Random initialization range.''' ) parser.add_argument( '''--fp16''' , action='''store_true''' , help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' , ) parser.add_argument( '''--fp16_opt_level''' , type=__lowerCAmelCase , default='''O1''' , help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) , ) parser.add_argument('''--n_gpu''' , type=__lowerCAmelCase , default=1 , help='''Number of GPUs in the node.''' ) parser.add_argument('''--local_rank''' , type=__lowerCAmelCase , default=-1 , help='''Distributed training - Local rank''' ) parser.add_argument('''--seed''' , type=__lowerCAmelCase , default=56 , help='''Random seed''' ) parser.add_argument('''--log_interval''' , type=__lowerCAmelCase , default=500 , help='''Tensorboard logging interval.''' ) parser.add_argument('''--checkpoint_interval''' , type=__lowerCAmelCase , default=4000 , help='''Checkpoint interval.''' ) __lowerCamelCase = parser.parse_args() sanity_checks(__lowerCAmelCase ) # ARGS # init_gpu_params(__lowerCAmelCase ) set_seed(__lowerCAmelCase ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f'''Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite''' ''' itUse `--force` if you want to overwrite it''' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f'''Experiment will be dumped and logged in {args.dump_path}''' ) # SAVE PARAMS # logger.info(f'''Param: {args}''' ) with open(os.path.join(args.dump_path , '''parameters.json''' ) , '''w''' ) as f: json.dump(vars(__lowerCAmelCase ) , __lowerCAmelCase , indent=4 ) git_log(args.dump_path ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = MODEL_CLASSES[args.student_type] __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = MODEL_CLASSES[args.teacher_type] # TOKENIZER # __lowerCamelCase = teacher_tokenizer_class.from_pretrained(args.teacher_name ) __lowerCamelCase = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): __lowerCamelCase = tokenizer.all_special_tokens.index(__lowerCAmelCase ) __lowerCamelCase = tokenizer.all_special_ids[idx] logger.info(f'''Special tokens {special_tok_ids}''' ) __lowerCamelCase = special_tok_ids __lowerCamelCase = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f'''Loading data from {args.data_file}''' ) with open(args.data_file , '''rb''' ) as fp: __lowerCamelCase = pickle.load(__lowerCAmelCase ) if args.mlm: logger.info(f'''Loading token counts from {args.token_counts} (already pre-computed)''' ) with open(args.token_counts , '''rb''' ) as fp: __lowerCamelCase = pickle.load(__lowerCAmelCase ) __lowerCamelCase = np.maximum(__lowerCAmelCase , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): __lowerCamelCase = 0.0 # do not predict special tokens __lowerCamelCase = torch.from_numpy(__lowerCAmelCase ) else: __lowerCamelCase = None __lowerCamelCase = LmSeqsDataset(params=__lowerCAmelCase , data=__lowerCAmelCase ) logger.info('''Data loader created.''' ) # STUDENT # logger.info(f'''Loading student config from {args.student_config}''' ) __lowerCamelCase = student_config_class.from_pretrained(args.student_config ) __lowerCamelCase = True if args.student_pretrained_weights is not None: logger.info(f'''Loading pretrained weights from {args.student_pretrained_weights}''' ) __lowerCamelCase = student_model_class.from_pretrained(args.student_pretrained_weights , config=__lowerCAmelCase ) else: __lowerCamelCase = student_model_class(__lowerCAmelCase ) if args.n_gpu > 0: student.to(f'''cuda:{args.local_rank}''' ) logger.info('''Student loaded.''' ) # TEACHER # __lowerCamelCase = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=__lowerCAmelCase ) if args.n_gpu > 0: teacher.to(f'''cuda:{args.local_rank}''' ) logger.info(f'''Teacher loaded from {args.teacher_name}.''' ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(__lowerCAmelCase , __lowerCAmelCase ) if args.freeze_token_type_embds: freeze_token_type_embeddings(__lowerCAmelCase , __lowerCAmelCase ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() __lowerCamelCase = Distiller( params=__lowerCAmelCase , dataset=__lowerCAmelCase , token_probs=__lowerCAmelCase , student=__lowerCAmelCase , teacher=__lowerCAmelCase ) distiller.train() logger.info('''Let\'s go get some drinks.''' ) if __name__ == "__main__": main()
339
from collections import namedtuple import requests from lxml import html # type: ignore SCREAMING_SNAKE_CASE__ : List[Any] = namedtuple("covid_data", "cases deaths recovered") def __magic_name__ ( __lowerCAmelCase : str = "https://www.worldometers.info/coronavirus/" ) -> covid_data: __lowerCamelCase = '''//div[@class = "maincounter-number"]/span/text()''' return covid_data(*html.fromstring(requests.get(__lowerCAmelCase ).content ).xpath(__lowerCAmelCase ) ) SCREAMING_SNAKE_CASE__ : List[str] = "Total COVID-19 cases in the world: {}\nTotal deaths due to COVID-19 in the world: {}\nTotal COVID-19 patients recovered in the world: {}" print(fmt.format(*covid_stats()))
339
1
def _A ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : str ): """simple docstring""" a__ : int =len(SCREAMING_SNAKE_CASE ) a__ : int =len(SCREAMING_SNAKE_CASE ) a__ : int =( first_str_length if first_str_length > second_str_length else second_str_length ) a__ : list =[] for char_count in range(SCREAMING_SNAKE_CASE ): if char_count < first_str_length: output_list.append(first_str[char_count] ) if char_count < second_str_length: output_list.append(second_str[char_count] ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(alternative_string_arrange("""AB""", """XYZ"""), end=""" """)
95
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class UpperCAmelCase ( A_ ): A__ : jnp.ndarray @flax_register_to_config class UpperCAmelCase ( nn.Module ,A_ ,A_ ): A__ : int = 32 A__ : int = 4 A__ : int = 4 A__ : Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) A__ : Tuple[str] = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") A__ : Union[bool, Tuple[bool]] = False A__ : Tuple[int] = (3_20, 6_40, 12_80, 12_80) A__ : int = 2 A__ : Union[int, Tuple[int]] = 8 A__ : Optional[Union[int, Tuple[int]]] = None A__ : int = 12_80 A__ : float = 0.0 A__ : bool = False A__ : jnp.dtype = jnp.floataa A__ : bool = True A__ : int = 0 A__ : bool = False def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : jax.random.KeyArray ) -> FrozenDict: '''simple docstring''' snake_case : Dict = (1, self.in_channels, self.sample_size, self.sample_size) snake_case : Any = jnp.zeros(snake_case__ , dtype=jnp.floataa ) snake_case : List[str] = jnp.ones((1,) , dtype=jnp.intaa ) snake_case : str = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) snake_case , snake_case : Optional[int] = jax.random.split(snake_case__ ) snake_case : Union[str, Any] = {"params": params_rng, "dropout": dropout_rng} return self.init(snake_case__ , snake_case__ , snake_case__ , snake_case__ )["params"] def _SCREAMING_SNAKE_CASE (self : str ) -> Tuple: '''simple docstring''' snake_case : str = self.block_out_channels snake_case : Optional[Any] = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( "At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19." ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. snake_case : Tuple = self.num_attention_heads or self.attention_head_dim # input snake_case : Tuple = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time snake_case : Union[str, Any] = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) snake_case : Dict = FlaxTimestepEmbedding(snake_case__ , dtype=self.dtype ) snake_case : List[str] = self.only_cross_attention if isinstance(snake_case__ , snake_case__ ): snake_case : List[Any] = (only_cross_attention,) * len(self.down_block_types ) if isinstance(snake_case__ , snake_case__ ): snake_case : List[Any] = (num_attention_heads,) * len(self.down_block_types ) # down snake_case : List[Any] = [] snake_case : Optional[int] = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): snake_case : List[Any] = output_channel snake_case : Dict = block_out_channels[i] snake_case : Optional[Any] = i == len(snake_case__ ) - 1 if down_block_type == "CrossAttnDownBlock2D": snake_case : List[Any] = FlaxCrossAttnDownBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: snake_case : Union[str, Any] = FlaxDownBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(snake_case__ ) snake_case : Dict = down_blocks # mid snake_case : Optional[int] = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up snake_case : Optional[Any] = [] snake_case : Optional[int] = list(reversed(snake_case__ ) ) snake_case : Dict = list(reversed(snake_case__ ) ) snake_case : Tuple = list(reversed(snake_case__ ) ) snake_case : Optional[Any] = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): snake_case : Optional[int] = output_channel snake_case : List[Any] = reversed_block_out_channels[i] snake_case : Union[str, Any] = reversed_block_out_channels[min(i + 1 , len(snake_case__ ) - 1 )] snake_case : int = i == len(snake_case__ ) - 1 if up_block_type == "CrossAttnUpBlock2D": snake_case : Any = FlaxCrossAttnUpBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , prev_output_channel=snake_case__ , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: snake_case : Optional[int] = FlaxUpBlockaD( in_channels=snake_case__ , out_channels=snake_case__ , prev_output_channel=snake_case__ , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(snake_case__ ) snake_case : Optional[int] = output_channel snake_case : Tuple = up_blocks # out snake_case : Optional[int] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) snake_case : List[str] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__(self : Dict , snake_case__ : Dict , snake_case__ : Dict , snake_case__ : Optional[int] , snake_case__ : Tuple=None , snake_case__ : Union[str, Any]=None , snake_case__ : bool = True , snake_case__ : bool = False , ) -> Union[FlaxUNetaDConditionOutput, Tuple]: '''simple docstring''' if not isinstance(snake_case__ , jnp.ndarray ): snake_case : List[Any] = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(snake_case__ , jnp.ndarray ) and len(timesteps.shape ) == 0: snake_case : Any = timesteps.astype(dtype=jnp.floataa ) snake_case : int = jnp.expand_dims(snake_case__ , 0 ) snake_case : str = self.time_proj(snake_case__ ) snake_case : str = self.time_embedding(snake_case__ ) # 2. pre-process snake_case : int = jnp.transpose(snake_case__ , (0, 2, 3, 1) ) snake_case : List[Any] = self.conv_in(snake_case__ ) # 3. down snake_case : Optional[int] = (sample,) for down_block in self.down_blocks: if isinstance(snake_case__ , snake_case__ ): snake_case , snake_case : List[Any] = down_block(snake_case__ , snake_case__ , snake_case__ , deterministic=not train ) else: snake_case , snake_case : str = down_block(snake_case__ , snake_case__ , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: snake_case : Tuple = () for down_block_res_sample, down_block_additional_residual in zip( snake_case__ , snake_case__ ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) snake_case : Optional[int] = new_down_block_res_samples # 4. mid snake_case : Optional[int] = self.mid_block(snake_case__ , snake_case__ , snake_case__ , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: snake_case : int = down_block_res_samples[-(self.layers_per_block + 1) :] snake_case : Optional[Any] = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(snake_case__ , snake_case__ ): snake_case : Optional[Any] = up_block( snake_case__ , temb=snake_case__ , encoder_hidden_states=snake_case__ , res_hidden_states_tuple=snake_case__ , deterministic=not train , ) else: snake_case : Dict = up_block(snake_case__ , temb=snake_case__ , res_hidden_states_tuple=snake_case__ , deterministic=not train ) # 6. post-process snake_case : List[str] = self.conv_norm_out(snake_case__ ) snake_case : Any = nn.silu(snake_case__ ) snake_case : Optional[int] = self.conv_out(snake_case__ ) snake_case : Union[str, Any] = jnp.transpose(snake_case__ , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=snake_case__ )
59
0
"""simple docstring""" import copy import os import cva import numpy as np from matplotlib import pyplot as plt class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self): __SCREAMING_SNAKE_CASE = """""" __SCREAMING_SNAKE_CASE = """""" __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 2_5_6 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = cva.imread(lowerCAmelCase__ , 0) __SCREAMING_SNAKE_CASE = copy.deepcopy(self.img) __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] , label="""x""") __SCREAMING_SNAKE_CASE = np.sum(lowerCAmelCase__) for i in range(len(lowerCAmelCase__)): __SCREAMING_SNAKE_CASE = x[i] / self.k self.sk += prk __SCREAMING_SNAKE_CASE = (self.L - 1) * self.sk if self.rem != 0: __SCREAMING_SNAKE_CASE = int(last % last) __SCREAMING_SNAKE_CASE = int(last + 1 if self.rem >= 0.5 else last) self.last_list.append(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = int(np.ma.count(self.img) / self.img[1].size) __SCREAMING_SNAKE_CASE = self.img[1].size for i in range(self.number_of_cols): for j in range(self.number_of_rows): __SCREAMING_SNAKE_CASE = self.img[j][i] if num != self.last_list[num]: __SCREAMING_SNAKE_CASE = self.last_list[num] cva.imwrite("""output_data/output.jpg""" , self.img) def snake_case_ ( self): plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6]) def snake_case_ ( self): cva.imshow("""Output-Image""" , self.img) cva.imshow("""Input-Image""" , self.original_image) cva.waitKey(5_0_0_0) cva.destroyAllWindows() if __name__ == "__main__": __magic_name__ = os.path.join(os.path.basename(__file__), "image_data/input.jpg") __magic_name__ = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
255
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def _lowerCAmelCase ( ): __SCREAMING_SNAKE_CASE = HfArgumentParser(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = parser.parse_args_into_dataclasses()[0] __SCREAMING_SNAKE_CASE = TensorFlowBenchmark(args=UpperCamelCase_ ) try: __SCREAMING_SNAKE_CASE = parser.parse_args_into_dataclasses()[0] except ValueError as e: __SCREAMING_SNAKE_CASE = """Arg --no_{0} is no longer used, please use --no-{0} instead.""" __SCREAMING_SNAKE_CASE = """ """.join(str(UpperCamelCase_ ).split(""" """ )[:-1] ) __SCREAMING_SNAKE_CASE = """""" __SCREAMING_SNAKE_CASE = eval(str(UpperCamelCase_ ).split(""" """ )[-1] ) __SCREAMING_SNAKE_CASE = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCamelCase_ ) if len(UpperCamelCase_ ) > 0: __SCREAMING_SNAKE_CASE = full_error_msg + begin_error_msg + str(UpperCamelCase_ ) raise ValueError(UpperCamelCase_ ) benchmark.run() if __name__ == "__main__": main()
255
1