code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging snake_case__ : Optional[int] = logging.get_logger(__name__) def _snake_case (__lowercase , __lowercase , __lowercase): UpperCamelCase_ = nn.ModuleList([src_layers[i] for i in layers_to_copy]) assert len(__lowercase) == len(__lowercase), f"""{len(__lowercase)} != {len(__lowercase)}""" dest_layers.load_state_dict(layers_to_copy.state_dict()) snake_case__ : Tuple = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 1_2: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 1_1], 4: [0, 4, 8, 1_1], 6: [0, 2, 4, 7, 9, 1_1], 9: [0, 1, 2, 4, 5, 7, 9, 1_0, 1_1], 1_2: list(range(1_2)), }, 1_6: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 1_5], 3: [0, 8, 1_5], 4: [0, 5, 1_0, 1_5], 6: [0, 3, 6, 9, 1_2, 1_5], 8: [0, 2, 4, 6, 8, 1_0, 1_2, 1_5], 9: [0, 1, 3, 5, 7, 9, 1_1, 1_3, 1_5], 1_2: [0, 1, 2, 3, 4, 5, 6, 7, 9, 1_1, 1_3, 1_5], 1_6: list(range(1_6)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } snake_case__ : Any = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 1_2: {1: [1_1], 2: [5, 1_1], 3: [3, 7, 1_1], 6: [1, 3, 5, 8, 1_0, 1_1]}, 1_6: {1: [1_5], 4: [4, 9, 1_2, 1_5], 8: [1, 3, 5, 7, 9, 1_1, 1_3, 1_5]}, } def _snake_case (__lowercase , __lowercase): try: UpperCamelCase_ = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( f"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" f""" {n_student}""") return list(range(__lowercase)) def _snake_case (__lowercase , __lowercase): if n_student > n_teacher: raise ValueError(f"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""") elif n_teacher == n_student: return list(range(__lowercase)) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _snake_case (__lowercase , __lowercase = "student" , __lowercase = None , __lowercase = None , __lowercase=False , __lowercase=None , __lowercase=None , **__lowercase , ): UpperCamelCase_ = 'encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.' assert (e is not None) or (d is not None), _msg if isinstance(__lowercase , __lowercase): AutoTokenizer.from_pretrained(__lowercase).save_pretrained(__lowercase) # purely for convenience UpperCamelCase_ = AutoModelForSeqaSeqLM.from_pretrained(__lowercase).eval() else: assert isinstance(__lowercase , __lowercase), f"""teacher must be a model or string got type {type(__lowercase)}""" UpperCamelCase_ = teacher.config.to_diff_dict() try: UpperCamelCase_ , UpperCamelCase_ = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: UpperCamelCase_ = teacher_e if d is None: UpperCamelCase_ = teacher_d init_kwargs.update({'encoder_layers': e, 'decoder_layers': d}) except AttributeError: # T5 if hasattr(teacher.config , 'num_encoder_layers'): UpperCamelCase_ , UpperCamelCase_ = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: UpperCamelCase_ , UpperCamelCase_ = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: UpperCamelCase_ = teacher_e if d is None: UpperCamelCase_ = teacher_d if hasattr(teacher.config , 'num_encoder_layers'): init_kwargs.update({'num_encoder_layers': e, 'num_decoder_layers': d}) else: init_kwargs.update({'num_layers': e, 'num_decoder_layers': d}) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(__lowercase) # Copy weights UpperCamelCase_ = teacher.config_class(**__lowercase) UpperCamelCase_ = AutoModelForSeqaSeqLM.from_config(__lowercase) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. UpperCamelCase_ = student.load_state_dict(teacher.state_dict() , strict=__lowercase) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save UpperCamelCase_ , UpperCamelCase_ = list(range(__lowercase)), list(range(__lowercase)) logger.info( f"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" f""" {save_path}""") student.save_pretrained(__lowercase) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: UpperCamelCase_ = pick_layers_to_copy(__lowercase , __lowercase) if d_layers_to_copy is None: UpperCamelCase_ = pick_layers_to_copy(__lowercase , __lowercase) try: if hasattr( __lowercase , 'prophetnet'): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , __lowercase) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , __lowercase) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , __lowercase) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , __lowercase) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , __lowercase) copy_layers(teacher.decoder.block , student.decoder.block , __lowercase) logger.info( f"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""") UpperCamelCase_ = { 'teacher_type': teacher.config.model_type, 'copied_encoder_layers': e_layers_to_copy, 'copied_decoder_layers': d_layers_to_copy, } student.save_pretrained(__lowercase) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
23
from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class _a ( UpperCAmelCase__ ): """simple docstring""" @slow @require_torch def _UpperCAmelCase ( self ) -> List[Any]: UpperCamelCase_ = EncoderDecoderModel.from_encoder_decoder_pretrained('prajjwal1/bert-tiny' , 'prajjwal1/bert-tiny' ) UpperCamelCase_ = BertTokenizer.from_pretrained('bert-base-uncased' ) UpperCamelCase_ = bertabert.config.encoder.vocab_size UpperCamelCase_ = tokenizer.sep_token_id UpperCamelCase_ = tokenizer.cls_token_id UpperCamelCase_ = 128 UpperCamelCase_ = datasets.load_dataset('cnn_dailymail' , '3.0.0' , split='train[:1%]' ) UpperCamelCase_ = datasets.load_dataset('cnn_dailymail' , '3.0.0' , split='validation[:1%]' ) UpperCamelCase_ = train_dataset.select(range(32 ) ) UpperCamelCase_ = val_dataset.select(range(16 ) ) UpperCamelCase_ = 4 def _map_to_encoder_decoder_inputs(_UpperCAmelCase ): # Tokenizer will automatically set [BOS] <text> [EOS] UpperCamelCase_ = tokenizer(batch['article'] , padding='max_length' , truncation=_UpperCAmelCase , max_length=512 ) UpperCamelCase_ = tokenizer(batch['highlights'] , padding='max_length' , truncation=_UpperCAmelCase , max_length=128 ) UpperCamelCase_ = inputs.input_ids UpperCamelCase_ = inputs.attention_mask UpperCamelCase_ = outputs.input_ids UpperCamelCase_ = outputs.input_ids.copy() UpperCamelCase_ = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch['labels'] ] UpperCamelCase_ = outputs.attention_mask assert all(len(_UpperCAmelCase ) == 512 for x in inputs.input_ids ) assert all(len(_UpperCAmelCase ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(_UpperCAmelCase ): UpperCamelCase_ = pred.label_ids UpperCamelCase_ = pred.predictions # all unnecessary tokens are removed UpperCamelCase_ = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) UpperCamelCase_ = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) UpperCamelCase_ = sum([int(pred_str[i] == label_str[i] ) for i in range(len(_UpperCAmelCase ) )] ) / len(_UpperCAmelCase ) return {"accuracy": accuracy} # map train dataset UpperCamelCase_ = train_dataset.map( _map_to_encoder_decoder_inputs , batched=_UpperCAmelCase , batch_size=_UpperCAmelCase , remove_columns=['article', 'highlights'] , ) train_dataset.set_format( type='torch' , columns=['input_ids', 'attention_mask', 'decoder_input_ids', 'decoder_attention_mask', 'labels'] , ) # same for validation dataset UpperCamelCase_ = val_dataset.map( _map_to_encoder_decoder_inputs , batched=_UpperCAmelCase , batch_size=_UpperCAmelCase , remove_columns=['article', 'highlights'] , ) val_dataset.set_format( type='torch' , columns=['input_ids', 'attention_mask', 'decoder_input_ids', 'decoder_attention_mask', 'labels'] , ) UpperCamelCase_ = self.get_auto_remove_tmp_dir() UpperCamelCase_ = SeqaSeqTrainingArguments( output_dir=_UpperCAmelCase , per_device_train_batch_size=_UpperCAmelCase , per_device_eval_batch_size=_UpperCAmelCase , predict_with_generate=_UpperCAmelCase , evaluation_strategy='steps' , do_train=_UpperCAmelCase , do_eval=_UpperCAmelCase , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer UpperCamelCase_ = SeqaSeqTrainer( model=_UpperCAmelCase , args=_UpperCAmelCase , compute_metrics=_compute_metrics , train_dataset=_UpperCAmelCase , eval_dataset=_UpperCAmelCase , tokenizer=_UpperCAmelCase , ) # start training trainer.train()
23
1
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __A : def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=32 , UpperCAmelCase_=3 , UpperCAmelCase_=4 , UpperCAmelCase_=[10, 20, 30, 40] , UpperCAmelCase_=[2, 2, 3, 2] , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=37 , UpperCAmelCase_="gelu" , UpperCAmelCase_=10 , UpperCAmelCase_=0.0_2 , UpperCAmelCase_=["stage2", "stage3", "stage4"] , UpperCAmelCase_=3 , UpperCAmelCase_=None , ): lowerCamelCase =parent lowerCamelCase =batch_size lowerCamelCase =image_size lowerCamelCase =num_channels lowerCamelCase =num_stages lowerCamelCase =hidden_sizes lowerCamelCase =depths lowerCamelCase =is_training lowerCamelCase =use_labels lowerCamelCase =intermediate_size lowerCamelCase =hidden_act lowerCamelCase =type_sequence_label_size lowerCamelCase =initializer_range lowerCamelCase =out_features lowerCamelCase =num_labels lowerCamelCase =scope lowerCamelCase =num_stages def _snake_case ( self ): lowerCamelCase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase =None if self.use_labels: lowerCamelCase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase =self.get_config() return config, pixel_values, labels def _snake_case ( self ): return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def _snake_case ( self ): return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=512 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=UpperCamelCase_ , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=256 , auxiliary_num_convs=1 , auxiliary_concat_input=UpperCamelCase_ , loss_ignore_index=255 , num_labels=self.num_labels , ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =UperNetForSemanticSegmentation(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() lowerCamelCase =model(UpperCamelCase_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def _snake_case ( self ): lowerCamelCase =self.prepare_config_and_inputs() ( lowerCamelCase ) =config_and_inputs lowerCamelCase ={'pixel_values': pixel_values} return config, inputs_dict @require_torch class __A ( a , a , unittest.TestCase ): __A = (UperNetForSemanticSegmentation,) if is_torch_available() else () __A = {"""image-segmentation""": UperNetForSemanticSegmentation} if is_torch_available() else {} __A = False __A = False __A = False __A = False __A = False __A = False def _snake_case ( self ): lowerCamelCase =UperNetModelTester(self ) lowerCamelCase =ConfigTester(self , config_class=UpperCamelCase_ , has_text_modality=UpperCamelCase_ , hidden_size=37 ) def _snake_case ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _snake_case ( self ): return def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase =model_class(UpperCamelCase_ ) lowerCamelCase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase =[*signature.parameters.keys()] lowerCamelCase =['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCamelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCamelCase_ ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def _snake_case ( self ): pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def _snake_case ( self ): pass @unittest.skip(reason="""UperNet does not have a base model""" ) def _snake_case ( self ): pass @unittest.skip(reason="""UperNet does not have a base model""" ) def _snake_case ( self ): pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`""" ) def _snake_case ( self ): pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _snake_case ( self ): pass def _snake_case ( self ): def check_hidden_states_output(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =model_class(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() with torch.no_grad(): lowerCamelCase =model(**self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) ) lowerCamelCase =outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCamelCase =self.model_tester.num_stages self.assertEqual(len(UpperCamelCase_ ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCamelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase =True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase =True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase =_config_zero_init(UpperCamelCase_ ) lowerCamelCase =_config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: lowerCamelCase =model_class(config=UpperCamelCase_ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def _snake_case ( self ): pass @slow def _snake_case ( self ): for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase =UperNetForSemanticSegmentation.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) def _lowercase ( ) -> int: lowerCamelCase =hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""" , repo_type="""dataset""" , filename="""ADE_val_00000001.jpg""" ) lowerCamelCase =Image.open(_lowercase ).convert("""RGB""" ) return image @require_torch @require_vision @slow class __A ( unittest.TestCase ): def _snake_case ( self ): lowerCamelCase =AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) lowerCamelCase =UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(UpperCamelCase_ ) lowerCamelCase =prepare_img() lowerCamelCase =processor(images=UpperCamelCase_ , return_tensors="""pt""" ).to(UpperCamelCase_ ) with torch.no_grad(): lowerCamelCase =model(**UpperCamelCase_ ) lowerCamelCase =torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCamelCase_ ) lowerCamelCase =torch.tensor( [[-7.5_9_5_8, -7.5_9_5_8, -7.4_3_0_2], [-7.5_9_5_8, -7.5_9_5_8, -7.4_3_0_2], [-7.4_7_9_7, -7.4_7_9_7, -7.3_0_6_8]] ).to(UpperCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCamelCase_ , atol=1E-4 ) ) def _snake_case ( self ): lowerCamelCase =AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) lowerCamelCase =UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(UpperCamelCase_ ) lowerCamelCase =prepare_img() lowerCamelCase =processor(images=UpperCamelCase_ , return_tensors="""pt""" ).to(UpperCamelCase_ ) with torch.no_grad(): lowerCamelCase =model(**UpperCamelCase_ ) lowerCamelCase =torch.Size((1, model.config.num_labels, 512, 512) ) self.assertEqual(outputs.logits.shape , UpperCamelCase_ ) lowerCamelCase =torch.tensor( [[-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.7_7_4_6, -8.7_7_4_6, -8.6_1_3_0]] ).to(UpperCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCamelCase_ , atol=1E-4 ) )
704
import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint UpperCAmelCase__ : Optional[int] ={ '''169M''': 12, '''430M''': 24, '''1B5''': 24, '''3B''': 32, '''7B''': 32, '''14B''': 40, } UpperCAmelCase__ : str ={ '''169M''': 7_68, '''430M''': 10_24, '''1B5''': 20_48, '''3B''': 25_60, '''7B''': 40_96, '''14B''': 51_20, } def _lowercase ( _UpperCAmelCase ) -> Tuple: lowerCamelCase =list(state_dict.keys() ) for name in state_dict_keys: lowerCamelCase =state_dict.pop(_UpperCAmelCase ) # emb -> embedding if name.startswith("""emb.""" ): lowerCamelCase =name.replace("""emb.""" , """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): lowerCamelCase =name.replace("""blocks.0.ln0""" , """blocks.0.pre_ln""" ) # att -> attention lowerCamelCase =re.sub(r"""blocks\.(\d+)\.att""" , r"""blocks.\1.attention""" , _UpperCAmelCase ) # ffn -> feed_forward lowerCamelCase =re.sub(r"""blocks\.(\d+)\.ffn""" , r"""blocks.\1.feed_forward""" , _UpperCAmelCase ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): lowerCamelCase =name.replace(""".time_mix_k""" , """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): lowerCamelCase =name.replace(""".time_mix_v""" , """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): lowerCamelCase =name.replace(""".time_mix_r""" , """.time_mix_receptance""" ) if name != "head.weight": lowerCamelCase ="""rwkv.""" + name lowerCamelCase =weight return state_dict def _lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=False , _UpperCAmelCase=None ) -> Tuple: # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) lowerCamelCase =5_02_77 lowerCamelCase =AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: lowerCamelCase =PreTrainedTokenizerFast(tokenizer_file=_UpperCAmelCase ) lowerCamelCase =len(_UpperCAmelCase ) tokenizer.save_pretrained(_UpperCAmelCase ) # 2. Build the config lowerCamelCase =list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: lowerCamelCase =candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(F"""`size` should be one of {possible_sizes}, got {size}.""" ) lowerCamelCase =RwkvConfig( vocab_size=_UpperCAmelCase , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(_UpperCAmelCase ) # 3. Download model file then convert state_dict lowerCamelCase =hf_hub_download(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase =torch.load(_UpperCAmelCase , map_location="""cpu""" ) lowerCamelCase =convert_state_dict(_UpperCAmelCase ) # 4. Split in shards and save lowerCamelCase , lowerCamelCase =shard_checkpoint(_UpperCAmelCase ) for shard_file, shard in shards.items(): torch.save(_UpperCAmelCase , os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) if index is not None: lowerCamelCase =os.path.join(_UpperCAmelCase , _UpperCAmelCase ) # Save the index as well with open(_UpperCAmelCase , """w""" , encoding="""utf-8""" ) as f: lowerCamelCase =json.dumps(_UpperCAmelCase , indent=2 , sort_keys=_UpperCAmelCase ) + """\n""" f.write(_UpperCAmelCase ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) lowerCamelCase =list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: lowerCamelCase =torch.load(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) lowerCamelCase =AutoModelForCausalLM.from_pretrained(_UpperCAmelCase ) model.push_to_hub(_UpperCAmelCase , max_shard_size="""2GB""" ) tokenizer.push_to_hub(_UpperCAmelCase ) if __name__ == "__main__": UpperCAmelCase__ : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--repo_id''', default=None, type=str, required=True, help='''Repo ID from which to pull the checkpoint.''' ) parser.add_argument( '''--checkpoint_file''', default=None, type=str, required=True, help='''Name of the checkpoint file in the repo.''' ) parser.add_argument( '''--output_dir''', default=None, type=str, required=True, help='''Where to save the converted model.''' ) parser.add_argument( '''--tokenizer_file''', default=None, type=str, help='''Path to the tokenizer file to use (if not provided, only the model is converted).''', ) parser.add_argument( '''--size''', default=None, type=str, help='''Size of the model. Will be inferred from the `checkpoint_file` if not passed.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Push to the Hub the converted model.''', ) parser.add_argument( '''--model_name''', default=None, type=str, help='''Name of the pushed model on the Hub, including the username / organization.''', ) UpperCAmelCase__ : List[Any] =parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
269
0
"""simple docstring""" import string from math import logaa def _A (__a , __a ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = document.translate( str.maketrans('''''' , '''''' , string.punctuation ) ).replace('''\n''' , '''''' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = document_without_punctuation.split(''' ''' ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def _A (__a , __a ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE_ : str = corpus.lower().translate( str.maketrans('''''' , '''''' , string.punctuation ) ) # strip all punctuation and replace it with '' SCREAMING_SNAKE_CASE_ : str = corpus_without_punctuation.split('''\n''' ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = term.lower() return (len([doc for doc in docs if term in doc] ), len(_lowercase )) def _A (__a , __a , __a=False ) -> Optional[Any]: """simple docstring""" if smoothing: if n == 0: raise ValueError('''log10(0) is undefined.''' ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError('''df must be > 0''' ) elif n == 0: raise ValueError('''log10(0) is undefined.''' ) return round(logaa(n / df ) , 3 ) def _A (__a , __a ) -> Dict: """simple docstring""" return round(tf * idf , 3 )
512
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A = logging.get_logger(__name__) __A = {'vocab_file': 'spiece.model'} __A = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class SCREAMING_SNAKE_CASE ( snake_case ): """simple docstring""" def __init__( self: Any , __A: Tuple , __A: int=False , __A: Tuple=True , __A: Optional[Any]=False , __A: int="<s>" , __A: Union[str, Any]="</s>" , __A: Dict="<unk>" , __A: int="<sep>" , __A: Dict="<pad>" , __A: Union[str, Any]="<cls>" , __A: Optional[int]="<mask>" , __A: Optional[Any]=["<eop>", "<eod>"] , __A: Optional[Dict[str, Any]] = None , **__A: List[Any] , ) -> None: _A = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else mask_token _A = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__A , remove_space=__A , keep_accents=__A , bos_token=__A , eos_token=__A , unk_token=__A , sep_token=__A , pad_token=__A , cls_token=__A , mask_token=__A , additional_special_tokens=__A , sp_model_kwargs=self.sp_model_kwargs , **__A , ) _A = 3 _A = do_lower_case _A = remove_space _A = keep_accents _A = vocab_file _A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__A ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( '''You need to install jieba to use CpmTokenizer or CpmTokenizerFast. ''' '''See https://pypi.org/project/jieba/ for installation.''' ) _A = jieba _A = str.maketrans(''' \n''' , '''\u2582\u2583''' ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def __A ( self: Optional[Any] ) -> Optional[Any]: return len(self.sp_model ) def __A ( self: int ) -> Optional[Any]: _A = {self.convert_ids_to_tokens(__A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self: Optional[Any] ) -> int: _A = self.__dict__.copy() _A = None return state def __setstate__( self: List[Any] , __A: List[Any] ) -> str: _A = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _A = {} _A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __A ( self: int , __A: Dict ) -> Dict: if self.remove_space: _A = ''' '''.join(inputs.strip().split() ) else: _A = inputs _A = outputs.replace('''``''' , '''"''' ).replace('''\'\'''' , '''"''' ) if not self.keep_accents: _A = unicodedata.normalize('''NFKD''' , __A ) _A = ''''''.join([c for c in outputs if not unicodedata.combining(__A )] ) if self.do_lower_case: _A = outputs.lower() return outputs def __A ( self: List[Any] , __A: str ) -> List[str]: _A = self.preprocess_text(__A ) _A = self.sp_model.encode(__A , out_type=__A ) _A = [] for piece in pieces: if len(__A ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): _A = self.sp_model.EncodeAsPieces(piece[:-1].replace(__A , '''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: _A = cur_pieces[1:] else: _A = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__A ) else: new_pieces.append(__A ) return new_pieces def __A ( self: str , __A: List[Any] ) -> Any: return self.sp_model.PieceToId(__A ) def __A ( self: List[str] , __A: Union[str, Any] ) -> Tuple: return self.sp_model.IdToPiece(__A ) def __A ( self: List[str] , __A: Optional[Any] ) -> Dict: _A = ''''''.join(__A ).replace(__A , ''' ''' ).strip() return out_string def __A ( self: str , __A: List[int] , __A: Optional[List[int]] = None ) -> List[int]: _A = [self.sep_token_id] _A = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def __A ( self: Dict , __A: List[int] , __A: Optional[List[int]] = None , __A: bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__A , token_ids_a=__A , already_has_special_tokens=__A ) if token_ids_a is not None: return ([0] * len(__A )) + [1] + ([0] * len(__A )) + [1, 1] return ([0] * len(__A )) + [1, 1] def __A ( self: Optional[Any] , __A: List[int] , __A: Optional[List[int]] = None ) -> List[int]: _A = [self.sep_token_id] _A = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def __A ( self: str , __A: str , __A: Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__A ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _A = os.path.join( __A , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __A ) elif not os.path.isfile(self.vocab_file ): with open(__A , '''wb''' ) as fi: _A = self.sp_model.serialized_model_proto() fi.write(__A ) return (out_vocab_file,) def __A ( self: Tuple , *__A: str , **__A: List[Any] ) -> Any: _A = super()._decode(*__A , **__A ) _A = text.replace(''' ''' , '''''' ).replace('''\u2582''' , ''' ''' ).replace('''\u2583''' , '''\n''' ) return text
484
0
import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase : Optional[int] = logging.get_logger(__name__) __UpperCAmelCase : str = { 'kakaobrain/align-base': 'https://huggingface.co/kakaobrain/align-base/resolve/main/config.json', } class lowerCamelCase ( SCREAMING_SNAKE_CASE ): UpperCAmelCase : List[Any] = 'align_text_model' def __init__( self : str , __snake_case : str=30522 , __snake_case : str=768 , __snake_case : Any=12 , __snake_case : Tuple=12 , __snake_case : Tuple=3072 , __snake_case : Dict="gelu" , __snake_case : Optional[int]=0.1 , __snake_case : List[Any]=0.1 , __snake_case : List[str]=512 , __snake_case : List[str]=2 , __snake_case : Dict=0.02 , __snake_case : Optional[Any]=1E-1_2 , __snake_case : List[Any]=0 , __snake_case : Optional[Any]="absolute" , __snake_case : Tuple=True , **__snake_case : Any , ) -> List[Any]: super().__init__(**__snake_case ) _a : Dict = vocab_size _a : Optional[Any] = hidden_size _a : Optional[int] = num_hidden_layers _a : Any = num_attention_heads _a : Optional[int] = hidden_act _a : List[str] = intermediate_size _a : Optional[int] = hidden_dropout_prob _a : Optional[int] = attention_probs_dropout_prob _a : List[Any] = max_position_embeddings _a : Optional[Any] = type_vocab_size _a : Tuple = initializer_range _a : List[str] = layer_norm_eps _a : Optional[int] = position_embedding_type _a : Any = use_cache _a : Dict = pad_token_id @classmethod def snake_case_ ( cls : Union[str, Any] , __snake_case : Union[str, os.PathLike] , **__snake_case : Optional[int] ) -> "PretrainedConfig": cls._set_token_in_kwargs(__snake_case ) _a , _a : Dict = cls.get_config_dict(__snake_case , **__snake_case ) # get the text config dict if we are loading from AlignConfig if config_dict.get('''model_type''' ) == "align": _a : Dict = config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__snake_case , **__snake_case ) class lowerCamelCase ( SCREAMING_SNAKE_CASE ): UpperCAmelCase : Tuple = 'align_vision_model' def __init__( self : int , __snake_case : int = 3 , __snake_case : int = 600 , __snake_case : float = 2.0 , __snake_case : float = 3.1 , __snake_case : int = 8 , __snake_case : List[int] = [3, 3, 5, 3, 5, 5, 3] , __snake_case : List[int] = [32, 16, 24, 40, 80, 112, 192] , __snake_case : List[int] = [16, 24, 40, 80, 112, 192, 320] , __snake_case : List[int] = [] , __snake_case : List[int] = [1, 2, 2, 2, 1, 2, 1] , __snake_case : List[int] = [1, 2, 2, 3, 3, 4, 1] , __snake_case : List[int] = [1, 6, 6, 6, 6, 6, 6] , __snake_case : float = 0.25 , __snake_case : str = "swish" , __snake_case : int = 2560 , __snake_case : str = "mean" , __snake_case : float = 0.02 , __snake_case : float = 0.001 , __snake_case : float = 0.99 , __snake_case : float = 0.2 , **__snake_case : Optional[Any] , ) -> List[str]: super().__init__(**__snake_case ) _a : Dict = num_channels _a : int = image_size _a : int = width_coefficient _a : List[Any] = depth_coefficient _a : Union[str, Any] = depth_divisor _a : Optional[int] = kernel_sizes _a : Dict = in_channels _a : int = out_channels _a : List[str] = depthwise_padding _a : List[str] = strides _a : Optional[int] = num_block_repeats _a : List[str] = expand_ratios _a : Optional[Any] = squeeze_expansion_ratio _a : int = hidden_act _a : Optional[Any] = hidden_dim _a : int = pooling_type _a : Optional[Any] = initializer_range _a : List[str] = batch_norm_eps _a : List[str] = batch_norm_momentum _a : Any = drop_connect_rate _a : str = sum(__snake_case ) * 4 @classmethod def snake_case_ ( cls : Optional[Any] , __snake_case : Union[str, os.PathLike] , **__snake_case : Optional[Any] ) -> "PretrainedConfig": cls._set_token_in_kwargs(__snake_case ) _a , _a : List[str] = cls.get_config_dict(__snake_case , **__snake_case ) # get the vision config dict if we are loading from AlignConfig if config_dict.get('''model_type''' ) == "align": _a : Union[str, Any] = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__snake_case , **__snake_case ) class lowerCamelCase ( SCREAMING_SNAKE_CASE ): UpperCAmelCase : Optional[int] = 'align' UpperCAmelCase : str = True def __init__( self : Optional[Any] , __snake_case : Optional[Any]=None , __snake_case : Dict=None , __snake_case : Tuple=640 , __snake_case : Dict=1.0 , __snake_case : List[Any]=0.02 , **__snake_case : Any , ) -> List[str]: super().__init__(**__snake_case ) if text_config is None: _a : Dict = {} logger.info('''text_config is None. Initializing the AlignTextConfig with default values.''' ) if vision_config is None: _a : Dict = {} logger.info('''vision_config is None. Initializing the AlignVisionConfig with default values.''' ) _a : Optional[int] = AlignTextConfig(**__snake_case ) _a : Any = AlignVisionConfig(**__snake_case ) _a : int = projection_dim _a : List[str] = temperature_init_value _a : Dict = initializer_range @classmethod def snake_case_ ( cls : Any , __snake_case : AlignTextConfig , __snake_case : AlignVisionConfig , **__snake_case : Optional[int] ) -> Tuple: return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__snake_case ) def snake_case_ ( self : str ) -> Any: _a : Tuple = copy.deepcopy(self.__dict__ ) _a : Union[str, Any] = self.text_config.to_dict() _a : Optional[Any] = self.vision_config.to_dict() _a : List[Any] = self.__class__.model_type return output
249
from manim import * class lowerCamelCase ( SCREAMING_SNAKE_CASE ): def snake_case_ ( self : int ) -> Tuple: _a : Optional[int] = Rectangle(height=0.5 , width=0.5 ) _a : Dict = Rectangle(height=0.25 , width=0.25 ) _a : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _a : Optional[int] = [mem.copy() for i in range(6 )] _a : Tuple = [mem.copy() for i in range(6 )] _a : Dict = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) _a : Any = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) _a : str = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) _a : List[str] = Text('''CPU''' , font_size=24 ) _a : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__snake_case ) _a : Union[str, Any] = [mem.copy() for i in range(4 )] _a : Tuple = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) _a : Tuple = Text('''GPU''' , font_size=24 ) _a : List[Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) gpu.move_to([-1, -1, 0] ) self.add(__snake_case ) _a : Optional[int] = [mem.copy() for i in range(6 )] _a : Any = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) _a : Union[str, Any] = Text('''Model''' , font_size=24 ) _a : str = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) model.move_to([3, -1.0, 0] ) self.add(__snake_case ) _a : Optional[Any] = [] _a : Optional[Any] = [] _a : Any = [] for i, rect in enumerate(__snake_case ): rect.set_stroke(__snake_case ) _a : Optional[Any] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__snake_case , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__snake_case ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(model_cpu_arr[0] , direction=__snake_case , buff=0.0 ) else: cpu_target.next_to(model_cpu_arr[i - 1] , direction=__snake_case , buff=0.0 ) self.add(__snake_case ) model_cpu_arr.append(__snake_case ) self.add(*__snake_case , *__snake_case , *__snake_case ) _a : List[Any] = [mem.copy() for i in range(6 )] _a : str = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) _a : Union[str, Any] = Text('''Loaded Checkpoint''' , font_size=24 ) _a : Any = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) checkpoint.move_to([3, 0.5, 0] ) self.add(__snake_case ) _a : Dict = [] _a : Tuple = [] for i, rect in enumerate(__snake_case ): _a : str = fill.copy().set_fill(__snake_case , opacity=0.7 ) target.move_to(__snake_case ) ckpt_arr.append(__snake_case ) _a : Optional[int] = target.copy() if i < 5: cpu_target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.move_to(cpu_right_col_base[i - 5] ) ckpt_cpu_arr.append(__snake_case ) self.add(*__snake_case , *__snake_case ) _a : int = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _a : Union[str, Any] = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__snake_case , __snake_case ) _a : Any = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(__snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(__snake_case ) _a : str = MarkupText( f"""Based on the passed in configuration, weights are stored in\na variety of np.memmaps on disk or to a particular device.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) _a : Optional[Any] = [meta_mem.copy() for i in range(6 )] _a : Union[str, Any] = [meta_mem.copy() for i in range(6 )] _a : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) _a : Optional[int] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) _a : Optional[int] = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) _a : Dict = Text('''Disk''' , font_size=24 ) _a : List[str] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) disk.move_to([-4.0, -1.25, 0] ) self.play(Write(__snake_case , run_time=3 ) , Write(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) ) _a : List[Any] = [] for i, rect in enumerate(__snake_case ): _a : Dict = rect.copy() target.generate_target() target.target.move_to(disk_left_col_base[i] ).scale(0.5 ) animations.append(MoveToTarget(__snake_case , run_time=1.5 ) ) self.play(*__snake_case ) self.play(FadeOut(__snake_case ) ) _a : Optional[int] = MarkupText(f"""Then, the checkpoint is removed from memory\nthrough garbage collection.""" , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(__snake_case , run_time=3 ) ) self.play( FadeOut(__snake_case , __snake_case , *__snake_case , *__snake_case ) , ) self.wait()
249
1
'''simple docstring''' from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging SCREAMING_SNAKE_CASE__ : Union[str, Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class a__( snake_case__ ): def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) -> List[Any]: super().__init__() if hasattr(scheduler.config , 'steps_offset' ) and scheduler.config.steps_offset != 1: snake_case__ =( f"""The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`""" f""" should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure """ 'to update the config accordingly as leaving `steps_offset` might led to incorrect results' ' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,' ' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`' ' file' ) deprecate('steps_offset!=1' , '1.0.0' , _UpperCAmelCase , standard_warn=_UpperCAmelCase ) snake_case__ =dict(scheduler.config ) snake_case__ =1 snake_case__ =FrozenDict(_UpperCAmelCase ) if hasattr(scheduler.config , 'skip_prk_steps' ) and scheduler.config.skip_prk_steps is False: snake_case__ =( f"""The configuration file of this scheduler: {scheduler} has not set the configuration""" ' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make' ' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to' ' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face' ' Hub, it would be very nice if you could open a Pull request for the' ' `scheduler/scheduler_config.json` file' ) deprecate('skip_prk_steps not set' , '1.0.0' , _UpperCAmelCase , standard_warn=_UpperCAmelCase ) snake_case__ =dict(scheduler.config ) snake_case__ =True snake_case__ =FrozenDict(_UpperCAmelCase ) if safety_checker is None: logger.warning( f"""You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure""" ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .' ) self.register_modules( segmentation_model=_UpperCAmelCase , segmentation_processor=_UpperCAmelCase , vae=_UpperCAmelCase , text_encoder=_UpperCAmelCase , tokenizer=_UpperCAmelCase , unet=_UpperCAmelCase , scheduler=_UpperCAmelCase , safety_checker=_UpperCAmelCase , feature_extractor=_UpperCAmelCase , ) def _lowercase ( self , _UpperCAmelCase = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory snake_case__ =self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_UpperCAmelCase ) def _lowercase ( self ) -> Tuple: self.enable_attention_slicing(_UpperCAmelCase ) def _lowercase ( self ) -> Optional[int]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) snake_case__ =torch.device('cuda' ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(_UpperCAmelCase , _UpperCAmelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _lowercase ( self ) -> Union[str, Any]: if self.device != torch.device('meta' ) or not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_UpperCAmelCase , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 512 , _UpperCAmelCase = 512 , _UpperCAmelCase = 50 , _UpperCAmelCase = 7.5 , _UpperCAmelCase = None , _UpperCAmelCase = 1 , _UpperCAmelCase = 0.0 , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = "pil" , _UpperCAmelCase = True , _UpperCAmelCase = None , _UpperCAmelCase = 1 , **_UpperCAmelCase , ) -> Union[str, Any]: snake_case__ =self.segmentation_processor( text=[text] , images=[image] , padding='max_length' , return_tensors='pt' ).to(self.device ) snake_case__ =self.segmentation_model(**_UpperCAmelCase ) snake_case__ =torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() snake_case__ =self.numpy_to_pil(_UpperCAmelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask snake_case__ =StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=_UpperCAmelCase , image=_UpperCAmelCase , mask_image=_UpperCAmelCase , height=_UpperCAmelCase , width=_UpperCAmelCase , num_inference_steps=_UpperCAmelCase , guidance_scale=_UpperCAmelCase , negative_prompt=_UpperCAmelCase , num_images_per_prompt=_UpperCAmelCase , eta=_UpperCAmelCase , generator=_UpperCAmelCase , latents=_UpperCAmelCase , output_type=_UpperCAmelCase , return_dict=_UpperCAmelCase , callback=_UpperCAmelCase , callback_steps=_UpperCAmelCase , )
538
'''simple docstring''' import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class a__( snake_case__ , unittest.TestCase ): a_ : str = BertJapaneseTokenizer a_ : List[str] = False a_ : List[Any] = True def _lowercase ( self ) -> Union[str, Any]: super().setUp() snake_case__ =[ '[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは', '世界', '##世界', '、', '##、', '。', '##。', ] snake_case__ =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def _lowercase ( self , _UpperCAmelCase ) -> str: snake_case__ ='こんにちは、世界。 \nこんばんは、世界。' snake_case__ ='こんにちは 、 世界 。 こんばんは 、 世界 。' return input_text, output_text def _lowercase ( self , _UpperCAmelCase ) -> Any: snake_case__ , snake_case__ =self.get_input_output_texts(_UpperCAmelCase ) snake_case__ =tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) snake_case__ =tokenizer.decode(_UpperCAmelCase , clean_up_tokenization_spaces=_UpperCAmelCase ) return text, ids def _lowercase ( self ) -> Dict: pass # TODO add if relevant def _lowercase ( self ) -> List[str]: pass # TODO add if relevant def _lowercase ( self ) -> str: pass # TODO add if relevant def _lowercase ( self ) -> Any: snake_case__ =self.tokenizer_class(self.vocab_file ) snake_case__ =tokenizer.tokenize('こんにちは、世界。\nこんばんは、世界。' ) self.assertListEqual(_UpperCAmelCase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def _lowercase ( self ) -> List[Any]: snake_case__ =self.tokenizer_class(self.vocab_file , word_tokenizer_type='mecab' ) self.assertIsNotNone(_UpperCAmelCase ) snake_case__ ='こんにちは、世界。\nこんばんは、世界。' snake_case__ =tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) snake_case__ =os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(_UpperCAmelCase , 'wb' ) as handle: pickle.dump(_UpperCAmelCase , _UpperCAmelCase ) with open(_UpperCAmelCase , 'rb' ) as handle: snake_case__ =pickle.load(_UpperCAmelCase ) snake_case__ =tokenizer_new.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _lowercase ( self ) -> Any: snake_case__ =MecabTokenizer(mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def _lowercase ( self ) -> str: try: snake_case__ =MecabTokenizer(mecab_dic='unidic_lite' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def _lowercase ( self ) -> List[str]: try: snake_case__ =MecabTokenizer(mecab_dic='unidic' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def _lowercase ( self ) -> int: snake_case__ =MecabTokenizer(do_lower_case=_UpperCAmelCase , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iphone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def _lowercase ( self ) -> List[Any]: try: snake_case__ =MecabTokenizer( do_lower_case=_UpperCAmelCase , normalize_text=_UpperCAmelCase , mecab_option='-d /usr/local/lib/mecab/dic/jumandic' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '\u3000', '。'] , ) def _lowercase ( self ) -> List[Any]: snake_case__ =MecabTokenizer(normalize_text=_UpperCAmelCase , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', ' ', '。'] , ) @require_sudachi def _lowercase ( self ) -> List[Any]: snake_case__ =self.tokenizer_class(self.vocab_file , word_tokenizer_type='sudachi' ) self.assertIsNotNone(_UpperCAmelCase ) snake_case__ ='こんにちは、世界。\nこんばんは、世界。' snake_case__ =tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) snake_case__ =os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(_UpperCAmelCase , 'wb' ) as handle: pickle.dump(_UpperCAmelCase , _UpperCAmelCase ) with open(_UpperCAmelCase , 'rb' ) as handle: snake_case__ =pickle.load(_UpperCAmelCase ) snake_case__ =tokenizer_new.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @require_sudachi def _lowercase ( self ) -> Union[str, Any]: snake_case__ =SudachiTokenizer(sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def _lowercase ( self ) -> Optional[Any]: snake_case__ =SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='A' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国', '人', '参政', '権'] ) @require_sudachi def _lowercase ( self ) -> List[Any]: snake_case__ =SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='B' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人', '参政権'] ) @require_sudachi def _lowercase ( self ) -> Optional[Any]: snake_case__ =SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='C' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人参政権'] ) @require_sudachi def _lowercase ( self ) -> Tuple: snake_case__ =SudachiTokenizer(do_lower_case=_UpperCAmelCase , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iphone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def _lowercase ( self ) -> Dict: snake_case__ =SudachiTokenizer(normalize_text=_UpperCAmelCase , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', '\u3000', '。', ' ', ' '] , ) @require_sudachi def _lowercase ( self ) -> Optional[int]: snake_case__ =SudachiTokenizer(trim_whitespace=_UpperCAmelCase , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) @require_jumanpp def _lowercase ( self ) -> Optional[Any]: snake_case__ =self.tokenizer_class(self.vocab_file , word_tokenizer_type='jumanpp' ) self.assertIsNotNone(_UpperCAmelCase ) snake_case__ ='こんにちは、世界。\nこんばんは、世界。' snake_case__ =tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) snake_case__ =os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(_UpperCAmelCase , 'wb' ) as handle: pickle.dump(_UpperCAmelCase , _UpperCAmelCase ) with open(_UpperCAmelCase , 'rb' ) as handle: snake_case__ =pickle.load(_UpperCAmelCase ) snake_case__ =tokenizer_new.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @require_jumanpp def _lowercase ( self ) -> int: snake_case__ =JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def _lowercase ( self ) -> Tuple: snake_case__ =JumanppTokenizer(do_lower_case=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iphone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def _lowercase ( self ) -> Tuple: snake_case__ =JumanppTokenizer(normalize_text=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['ア', 'ッ', 'フ', '゚', 'ル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def _lowercase ( self ) -> int: snake_case__ =JumanppTokenizer(trim_whitespace=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '。'] , ) @require_jumanpp def _lowercase ( self ) -> str: snake_case__ =JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('ありがとうございますm(_ _)m見つけるのが大変です。' ) , ['ありがとう', 'ございます', 'm(_ _)m', '見つける', 'の', 'が', '大変です', '。'] , ) def _lowercase ( self ) -> Optional[Any]: snake_case__ =['[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは'] snake_case__ ={} for i, token in enumerate(_UpperCAmelCase ): snake_case__ =i snake_case__ =WordpieceTokenizer(vocab=_UpperCAmelCase , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こんにちは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは' ) , ['こん', '##ばんは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは こんばんにちは こんにちは' ) , ['こん', '##ばんは', '[UNK]', 'こんにちは'] ) def _lowercase ( self ) -> Tuple: snake_case__ =BertJapaneseTokenizer.from_pretrained('nlp-waseda/roberta-base-japanese-with-auto-jumanpp' ) snake_case__ =tokenizer.subword_tokenizer snake_case__ =subword_tokenizer.tokenize('国境 の 長い トンネル を 抜ける と 雪国 であった 。' ) self.assertListEqual(_UpperCAmelCase , ['▁国境', '▁の', '▁長い', '▁トンネル', '▁を', '▁抜ける', '▁と', '▁雪', '国', '▁であった', '▁。'] ) snake_case__ =subword_tokenizer.tokenize('こんばんは こんばん にち は こんにちは' ) self.assertListEqual(_UpperCAmelCase , ['▁こん', 'ばん', 'は', '▁こん', 'ばん', '▁に', 'ち', '▁は', '▁こんにちは'] ) def _lowercase ( self ) -> Optional[int]: snake_case__ =self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese' ) snake_case__ =tokenizer.encode('ありがとう。' , add_special_tokens=_UpperCAmelCase ) snake_case__ =tokenizer.encode('どういたしまして。' , add_special_tokens=_UpperCAmelCase ) snake_case__ =tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase ) snake_case__ =tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase , _UpperCAmelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class a__( snake_case__ , unittest.TestCase ): a_ : int = BertJapaneseTokenizer a_ : Optional[Any] = False def _lowercase ( self ) -> Any: super().setUp() snake_case__ =['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] snake_case__ =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def _lowercase ( self , **_UpperCAmelCase ) -> Union[str, Any]: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='character' , **_UpperCAmelCase ) def _lowercase ( self , _UpperCAmelCase ) -> Optional[Any]: snake_case__ ='こんにちは、世界。 \nこんばんは、世界。' snake_case__ ='こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。' return input_text, output_text def _lowercase ( self ) -> List[Any]: pass # TODO add if relevant def _lowercase ( self ) -> Union[str, Any]: pass # TODO add if relevant def _lowercase ( self ) -> Tuple: pass # TODO add if relevant def _lowercase ( self ) -> Any: snake_case__ =self.tokenizer_class(self.vocab_file , subword_tokenizer_type='character' ) snake_case__ =tokenizer.tokenize('こんにちは、世界。 \nこんばんは、世界。' ) self.assertListEqual( _UpperCAmelCase , ['こ', 'ん', 'に', 'ち', 'は', '、', '世', '界', '。', 'こ', 'ん', 'ば', 'ん', 'は', '、', '世', '界', '。'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def _lowercase ( self ) -> Optional[Any]: snake_case__ =['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] snake_case__ ={} for i, token in enumerate(_UpperCAmelCase ): snake_case__ =i snake_case__ =CharacterTokenizer(vocab=_UpperCAmelCase , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こ', 'ん', 'に', 'ち', 'は'] ) self.assertListEqual(tokenizer.tokenize('こんにちほ' ) , ['こ', 'ん', 'に', 'ち', '[UNK]'] ) def _lowercase ( self ) -> Any: snake_case__ =self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese-char' ) snake_case__ =tokenizer.encode('ありがとう。' , add_special_tokens=_UpperCAmelCase ) snake_case__ =tokenizer.encode('どういたしまして。' , add_special_tokens=_UpperCAmelCase ) snake_case__ =tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase ) snake_case__ =tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase , _UpperCAmelCase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class a__( unittest.TestCase ): def _lowercase ( self ) -> Optional[Any]: snake_case__ ='cl-tohoku/bert-base-japanese' snake_case__ =AutoTokenizer.from_pretrained(_UpperCAmelCase ) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) class a__( unittest.TestCase ): def _lowercase ( self ) -> int: snake_case__ ='cl-tohoku/bert-base-japanese' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertTokenizer.from_pretrained(_UpperCAmelCase ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) ) snake_case__ ='bert-base-cased' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertJapaneseTokenizer.from_pretrained(_UpperCAmelCase ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) )
538
1
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata __snake_case :int = '''''' if version.parse(importlib_metadata.version('''jiwer''')) < version.parse('''2.3.0'''): class _A ( tr.AbstractTransform ): def __init__( self : List[Any] , __SCREAMING_SNAKE_CASE : str = " "): '''simple docstring''' __a = sentence_delimiter def _lowerCamelCase ( self : Dict , __SCREAMING_SNAKE_CASE : str): '''simple docstring''' return list(__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : int , __SCREAMING_SNAKE_CASE : List[str]): '''simple docstring''' __a = [] for sent_idx, sentence in enumerate(__SCREAMING_SNAKE_CASE): chars.extend(self.process_string(__SCREAMING_SNAKE_CASE)) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(__SCREAMING_SNAKE_CASE) - 1: chars.append(self.sentence_delimiter) return chars __snake_case :Any = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: __snake_case :Optional[int] = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) __snake_case :Optional[int] = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' __snake_case :Tuple = '''\ Character error rate (CER) is a common metric of the performance of an automatic speech recognition system. CER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information. Character error rate can be computed as: CER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct characters, N is the number of characters in the reference (N=S+D+C). CER\'s output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the performance of the ASR system with a CER of 0 being a perfect score. ''' __snake_case :Tuple = ''' Computes CER score of transcribed segments against references. Args: references: list of references for each speech input. predictions: list of transcribtions to score. concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result. Returns: (float): the character error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> cer = datasets.load_metric("cer") >>> cer_score = cer.compute(predictions=predictions, references=references) >>> print(cer_score) 0.34146341463414637 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class _A ( datasets.Metric ): def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence'''), '''references''': datasets.Value('''string''' , id='''sequence'''), }) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', '''https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates''', ] , ) def _lowerCamelCase ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Dict=False): '''simple docstring''' if concatenate_texts: return jiwer.compute_measures( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , truth_transform=__SCREAMING_SNAKE_CASE , hypothesis_transform=__SCREAMING_SNAKE_CASE , )["wer"] __a = 0 __a = 0 for prediction, reference in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE): __a = jiwer.compute_measures( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , truth_transform=__SCREAMING_SNAKE_CASE , hypothesis_transform=__SCREAMING_SNAKE_CASE , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
60
from __future__ import annotations from random import random from typing import Generic, TypeVar __snake_case :Any = TypeVar('''KT''') __snake_case :List[str] = TypeVar('''VT''') class _A ( Generic[KT, VT] ): def __init__( self : Dict , __SCREAMING_SNAKE_CASE : KT | str = "root" , __SCREAMING_SNAKE_CASE : VT | None = None): '''simple docstring''' __a = key __a = value __a = [] def __repr__( self : Dict): '''simple docstring''' return F'Node({self.key}: {self.value})' @property def _lowerCamelCase ( self : Tuple): '''simple docstring''' return len(self.forward) class _A ( Generic[KT, VT] ): def __init__( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : float = 0.5 , __SCREAMING_SNAKE_CASE : int = 16): '''simple docstring''' __a = Node[KT, VT]() __a = 0 __a = p __a = max_level def __str__( self : Union[str, Any]): '''simple docstring''' __a = list(self) if len(__SCREAMING_SNAKE_CASE) == 0: return F'SkipList(level={self.level})' __a = max((len(str(__SCREAMING_SNAKE_CASE)) for item in items) , default=4) __a = max(__SCREAMING_SNAKE_CASE , 4) + 4 __a = self.head __a = [] __a = node.forward.copy() lines.append(F'[{node.key}]'.ljust(__SCREAMING_SNAKE_CASE , '''-''') + '''* ''' * len(__SCREAMING_SNAKE_CASE)) lines.append(''' ''' * label_size + '''| ''' * len(__SCREAMING_SNAKE_CASE)) while len(node.forward) != 0: __a = node.forward[0] lines.append( F'[{node.key}]'.ljust(__SCREAMING_SNAKE_CASE , '''-''') + ''' '''.join(str(n.key) if n.key == node.key else '''|''' for n in forwards)) lines.append(''' ''' * label_size + '''| ''' * len(__SCREAMING_SNAKE_CASE)) __a = node.forward lines.append('''None'''.ljust(__SCREAMING_SNAKE_CASE) + '''* ''' * len(__SCREAMING_SNAKE_CASE)) return F'SkipList(level={self.level})\n' + "\n".join(__SCREAMING_SNAKE_CASE) def __iter__( self : int): '''simple docstring''' __a = self.head while len(node.forward) != 0: yield node.forward[0].key __a = node.forward[0] def _lowerCamelCase ( self : List[Any]): '''simple docstring''' __a = 1 while random() < self.p and level < self.max_level: level += 1 return level def _lowerCamelCase ( self : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any]): '''simple docstring''' __a = [] __a = self.head for i in reversed(range(self.level)): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: __a = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(__SCREAMING_SNAKE_CASE) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def _lowerCamelCase ( self : Dict , __SCREAMING_SNAKE_CASE : KT): '''simple docstring''' __a , __a = self._locate_node(__SCREAMING_SNAKE_CASE) if node is not None: for i, update_node in enumerate(__SCREAMING_SNAKE_CASE): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: __a = node.forward[i] else: __a = update_node.forward[:i] def _lowerCamelCase ( self : List[str] , __SCREAMING_SNAKE_CASE : KT , __SCREAMING_SNAKE_CASE : VT): '''simple docstring''' __a , __a = self._locate_node(__SCREAMING_SNAKE_CASE) if node is not None: __a = value else: __a = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , __SCREAMING_SNAKE_CASE): update_vector.append(self.head) __a = level __a = Node(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE) for i, update_node in enumerate(update_vector[:level]): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i]) if update_node.level < i + 1: update_node.forward.append(__SCREAMING_SNAKE_CASE) else: __a = new_node def _lowerCamelCase ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : VT): '''simple docstring''' __a , __a = self._locate_node(__SCREAMING_SNAKE_CASE) if node is not None: return node.value return None def __snake_case ( ): __a = SkipList() skip_list.insert('''Key1''' , 3 ) skip_list.insert('''Key2''' , 12 ) skip_list.insert('''Key3''' , 41 ) skip_list.insert('''Key4''' , -19 ) __a = skip_list.head __a = {} while node.level != 0: __a = node.forward[0] __a = node.value assert len(_UpperCAmelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def __snake_case ( ): __a = SkipList() skip_list.insert('''Key1''' , 10 ) skip_list.insert('''Key1''' , 12 ) skip_list.insert('''Key5''' , 7 ) skip_list.insert('''Key7''' , 10 ) skip_list.insert('''Key10''' , 5 ) skip_list.insert('''Key7''' , 7 ) skip_list.insert('''Key5''' , 5 ) skip_list.insert('''Key10''' , 10 ) __a = skip_list.head __a = {} while node.level != 0: __a = node.forward[0] __a = node.value if len(_UpperCAmelCase ) != 4: print() assert len(_UpperCAmelCase ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def __snake_case ( ): __a = SkipList() assert skip_list.find('''Some key''' ) is None def __snake_case ( ): __a = SkipList() skip_list.insert('''Key2''' , 20 ) assert skip_list.find('''Key2''' ) == 20 skip_list.insert('''Some Key''' , 10 ) skip_list.insert('''Key2''' , 8 ) skip_list.insert('''V''' , 13 ) assert skip_list.find('''Y''' ) is None assert skip_list.find('''Key2''' ) == 8 assert skip_list.find('''Some Key''' ) == 10 assert skip_list.find('''V''' ) == 13 def __snake_case ( ): __a = SkipList() skip_list.delete('''Some key''' ) assert len(skip_list.head.forward ) == 0 def __snake_case ( ): __a = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 14 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''V''' ) skip_list.delete('''Key2''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''Key2''' ) is None def __snake_case ( ): __a = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 14 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''V''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) == 14 assert skip_list.find('''Key1''' ) == 12 assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''X''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) == 12 assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''Key1''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) is None assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''Key2''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) is None assert skip_list.find('''Key2''' ) is None def __snake_case ( ): __a = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 142 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''X''' ) def traverse_keys(_UpperCAmelCase ): yield node.key for forward_node in node.forward: yield from traverse_keys(_UpperCAmelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __snake_case ( ): def is_sorted(_UpperCAmelCase ): return all(next_item >= item for item, next_item in zip(_UpperCAmelCase , lst[1:] ) ) __a = SkipList() for i in range(10 ): skip_list.insert(_UpperCAmelCase , _UpperCAmelCase ) assert is_sorted(list(_UpperCAmelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_UpperCAmelCase ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(_UpperCAmelCase ) ) def __snake_case ( ): for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __snake_case ( ): __a = SkipList() skip_list.insert(2 , '''2''' ) skip_list.insert(4 , '''4''' ) skip_list.insert(6 , '''4''' ) skip_list.insert(4 , '''5''' ) skip_list.insert(8 , '''4''' ) skip_list.insert(9 , '''4''' ) skip_list.delete(4 ) print(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
60
1
import datasets SCREAMING_SNAKE_CASE__ : Any = """\ @InProceedings{conneau2018xnli, author = \"Conneau, Alexis and Rinott, Ruty and Lample, Guillaume and Williams, Adina and Bowman, Samuel R. and Schwenk, Holger and Stoyanov, Veselin\", title = \"XNLI: Evaluating Cross-lingual Sentence Representations\", booktitle = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing\", year = \"2018\", publisher = \"Association for Computational Linguistics\", location = \"Brussels, Belgium\", } """ SCREAMING_SNAKE_CASE__ : Optional[Any] = """\ XNLI is a subset of a few thousand examples from MNLI which has been translated into a 14 different languages (some low-ish resource). As with MNLI, the goal is to predict textual entailment (does sentence A imply/contradict/neither sentence B) and is a classification task (given two sentences, predict one of three labels). """ SCREAMING_SNAKE_CASE__ : Optional[int] = """ Computes XNLI score which is just simple accuracy. Args: predictions: Predicted labels. references: Ground truth labels. Returns: 'accuracy': accuracy Examples: >>> predictions = [0, 1] >>> references = [0, 1] >>> xnli_metric = datasets.load_metric(\"xnli\") >>> results = xnli_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} """ def __lowercase ( snake_case, snake_case ): """simple docstring""" return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): def A ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32''' ), '''references''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32''' ), } ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' , ) def A ( self , __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" return {"accuracy": simple_accuracy(__lowerCAmelCase , __lowerCAmelCase )}
0
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed snake_case__ : List[str] = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F'''{bindir}/../../examples/pytorch/translation'''): from run_translation import main # noqa set_seed(42) snake_case__ : Any = """sshleifer/student_marian_en_ro_6_1""" snake_case__ : Dict = """sshleifer/tiny-mbart""" @require_torch class _A ( _lowercase ): '''simple docstring''' def _snake_case ( self : Optional[Any] , lowerCamelCase : Dict=False , lowerCamelCase : Any=None , lowerCamelCase : int=True , lowerCamelCase : str=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Optional[Any]=True , ): '''simple docstring''' __lowercase = self.run_trainer( eval_steps=1 , max_len=12 , model_name=lowerCamelCase , num_train_epochs=1 , distributed=lowerCamelCase , extra_args_str=lowerCamelCase , predict_with_generate=lowerCamelCase , do_train=lowerCamelCase , do_eval=lowerCamelCase , do_predict=lowerCamelCase , ) __lowercase = TrainerState.load_from_json(os.path.join(lowerCamelCase , "trainer_state.json" ) ).log_history if not do_eval: return __lowercase = [log for log in logs if "eval_loss" in log.keys()] __lowercase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats __lowercase = eval_metrics[-1] assert isinstance(last_step_stats["eval_bleu"] , lowerCamelCase ) assert not math.isnan(float(last_step_stats["eval_loss"] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _snake_case ( self : Dict ): '''simple docstring''' self.run_seqaseq_quick() @require_torch_multi_gpu def _snake_case ( self : Optional[Any] ): '''simple docstring''' self.run_seqaseq_quick(distributed=lowerCamelCase ) @require_torch_multi_gpu def _snake_case ( self : Any ): '''simple docstring''' self.run_seqaseq_quick(distributed=lowerCamelCase ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def _snake_case ( self : Any ): '''simple docstring''' self.run_seqaseq_quick(distributed=lowerCamelCase , extra_args_str="--sharded_ddp simple" ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def _snake_case ( self : Any ): '''simple docstring''' self.run_seqaseq_quick(distributed=lowerCamelCase , extra_args_str="--sharded_ddp simple --fp16" ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def _snake_case ( self : str ): '''simple docstring''' self.run_seqaseq_quick(distributed=lowerCamelCase , extra_args_str="--sharded_ddp zero_dp_2" , predict_with_generate=lowerCamelCase ) @unittest.skip("Requires an update of the env running those tests" ) @require_torch_multi_gpu @require_fairscale def _snake_case ( self : List[Any] ): '''simple docstring''' self.run_seqaseq_quick( distributed=lowerCamelCase , extra_args_str="--sharded_ddp zero_dp_2 --fp16" , predict_with_generate=lowerCamelCase ) @require_apex @require_torch_gpu def _snake_case ( self : Any ): '''simple docstring''' self.run_seqaseq_quick(distributed=lowerCamelCase , extra_args_str="--fp16 --fp16_backend=apex" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=lowerCamelCase , extra_args_str="--fp16 --fp16_backend=apex" ) @parameterized.expand(["base", "low", "high", "mixed"] ) @require_torch_multi_gpu def _snake_case ( self : List[Any] , lowerCamelCase : Any ): '''simple docstring''' __lowercase = { # test with the default log_level - should be info and thus log info once "base": {"extra_args_str": "", "n_matches": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes "low": {"extra_args_str": "--log_level debug --log_level_replica debug", "n_matches": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica "high": {"extra_args_str": "--log_level error --log_level_replica debug", "n_matches": 1}, # test with high log_level and log_level_replica - should be quiet on all processes "mixed": {"extra_args_str": "--log_level error --log_level_replica error", "n_matches": 0}, } __lowercase = experiments[experiment_id] __lowercase = {"distributed": True, "predict_with_generate": False, "do_eval": False, "do_predict": False} __lowercase = "Running training" with CaptureStderr() as cl: self.run_seqaseq_quick(**lowerCamelCase , extra_args_str=data["extra_args_str"] ) __lowercase = len(re.findall(lowerCamelCase , cl.err ) ) self.assertEqual(lowerCamelCase , data["n_matches"] ) @slow def _snake_case ( self : Tuple ): '''simple docstring''' __lowercase = self.run_trainer( eval_steps=2 , max_len=128 , model_name=lowerCamelCase , learning_rate=3e-4 , num_train_epochs=10 , distributed=lowerCamelCase , ) # Check metrics __lowercase = TrainerState.load_from_json(os.path.join(lowerCamelCase , "trainer_state.json" ) ).log_history __lowercase = [log for log in logs if "eval_loss" in log.keys()] __lowercase = eval_metrics[0] __lowercase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["eval_bleu"] , lowerCamelCase ) # test if do_predict saves generations and metrics __lowercase = os.listdir(lowerCamelCase ) __lowercase = {os.path.basename(lowerCamelCase ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _snake_case ( self : Dict ): '''simple docstring''' from transformers.training_args import OptimizerNames def train_and_return_metrics(lowerCamelCase : str ) -> Tuple[int, float]: __lowercase = "--skip_memory_metrics 0" __lowercase = self.run_trainer( max_len=128 , model_name=lowerCamelCase , learning_rate=3e-4 , num_train_epochs=1 , optim=lowerCamelCase , distributed=lowerCamelCase , extra_args_str=lowerCamelCase , do_eval=lowerCamelCase , do_predict=lowerCamelCase , n_gpus_to_use=1 , ) # Check metrics __lowercase = TrainerState.load_from_json(Path(lowerCamelCase , "trainer_state.json" ) ).log_history __lowercase = int(logs[0]["train_mem_gpu_peaked_delta"] / 2**20 ) __lowercase = int(logs[0]["train_mem_gpu_alloc_delta"] / 2**20 ) __lowercase = logs[0]["train_loss"] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss __lowercase , __lowercase , __lowercase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) __lowercase , __lowercase , __lowercase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) __lowercase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb __lowercase = gpu_peak_mem_orig + gpu_alloc_mem_orig __lowercase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb __lowercase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings __lowercase = 120 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( lowerCamelCase , lowerCamelCase , "should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got" f""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" f""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( lowerCamelCase , lowerCamelCase , "should use ~150MB less total gpu memory with BNB, compared to without it for this model but got" f""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" f""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( lowerCamelCase , lowerCamelCase , f"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""" ) def _snake_case ( self : Dict , lowerCamelCase : int , lowerCamelCase : str , lowerCamelCase : int , lowerCamelCase : float = 3e-3 , lowerCamelCase : str = "adafactor" , lowerCamelCase : bool = False , lowerCamelCase : str = None , lowerCamelCase : int = 0 , lowerCamelCase : bool = True , lowerCamelCase : bool = True , lowerCamelCase : bool = True , lowerCamelCase : bool = True , lowerCamelCase : int = None , ): '''simple docstring''' __lowercase = self.test_file_dir / "../fixtures/tests_samples/wmt_en_ro" __lowercase = self.get_auto_remove_tmp_dir() __lowercase = f""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(lowerCamelCase )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(lowerCamelCase )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() __lowercase = f""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(lowerCamelCase )} """.split() __lowercase = "\n --do_predict\n ".split() __lowercase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += f"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: __lowercase = get_gpu_count() __lowercase = get_torch_dist_unique_port() __lowercase = f""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() __lowercase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowerCamelCase , env=self.get_env() ) else: __lowercase = ["run_translation.py"] + args with patch.object(lowerCamelCase , "argv" , lowerCamelCase ): main() return output_dir
402
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_ ( lowercase , lowercase , lowercase , unittest.TestCase ): __lowercase : Optional[int] = AltDiffusionPipeline __lowercase : Tuple = TEXT_TO_IMAGE_PARAMS __lowercase : Optional[Any] = TEXT_TO_IMAGE_BATCH_PARAMS __lowercase : Tuple = TEXT_TO_IMAGE_IMAGE_PARAMS __lowercase : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS def lowercase ( self ) -> List[str]: """simple docstring""" torch.manual_seed(0 ) _UpperCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) _UpperCamelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , clip_sample=lowerCamelCase_ , set_alpha_to_one=lowerCamelCase_ , ) torch.manual_seed(0 ) _UpperCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) _UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=50_02 , ) _UpperCamelCase = CLIPTextModel(lowerCamelCase_ ) _UpperCamelCase = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) _UpperCamelCase = 77 _UpperCamelCase = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def lowercase ( self , lowerCamelCase_ , lowerCamelCase_=0 ) -> Tuple: """simple docstring""" if str(lowerCamelCase_ ).startswith("mps" ): _UpperCamelCase = torch.manual_seed(lowerCamelCase_ ) else: _UpperCamelCase = torch.Generator(device=lowerCamelCase_ ).manual_seed(lowerCamelCase_ ) _UpperCamelCase = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def lowercase ( self ) -> str: """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def lowercase ( self ) -> Tuple: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowercase ( self ) -> Dict: """simple docstring""" _UpperCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCamelCase = self.get_dummy_components() torch.manual_seed(0 ) _UpperCamelCase = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=50_02 , ) # TODO: remove after fixing the non-deterministic text encoder _UpperCamelCase = RobertaSeriesModelWithTransformation(lowerCamelCase_ ) _UpperCamelCase = text_encoder _UpperCamelCase = AltDiffusionPipeline(**lowerCamelCase_ ) _UpperCamelCase = alt_pipe.to(lowerCamelCase_ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) _UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_ ) _UpperCamelCase = "A photo of an astronaut" _UpperCamelCase = alt_pipe(**lowerCamelCase_ ) _UpperCamelCase = output.images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowercase ( self ) -> int: """simple docstring""" _UpperCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCamelCase = self.get_dummy_components() _UpperCamelCase = PNDMScheduler(skip_prk_steps=lowerCamelCase_ ) torch.manual_seed(0 ) _UpperCamelCase = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=50_02 , ) # TODO: remove after fixing the non-deterministic text encoder _UpperCamelCase = RobertaSeriesModelWithTransformation(lowerCamelCase_ ) _UpperCamelCase = text_encoder _UpperCamelCase = AltDiffusionPipeline(**lowerCamelCase_ ) _UpperCamelCase = alt_pipe.to(lowerCamelCase_ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) _UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_ ) _UpperCamelCase = alt_pipe(**lowerCamelCase_ ) _UpperCamelCase = output.images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _UpperCamelCase = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): def lowercase ( self ) -> Optional[int]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self ) -> Optional[int]: """simple docstring""" _UpperCamelCase = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , safety_checker=lowerCamelCase_ ) _UpperCamelCase = alt_pipe.to(lowerCamelCase_ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) _UpperCamelCase = "A painting of a squirrel eating a burger" _UpperCamelCase = torch.manual_seed(0 ) _UpperCamelCase = alt_pipe([prompt] , generator=lowerCamelCase_ , guidance_scale=6.0 , num_inference_steps=20 , output_type="np" ) _UpperCamelCase = output.images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _UpperCamelCase = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowercase ( self ) -> Dict: """simple docstring""" _UpperCamelCase = DDIMScheduler.from_pretrained("BAAI/AltDiffusion" , subfolder="scheduler" ) _UpperCamelCase = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , scheduler=lowerCamelCase_ , safety_checker=lowerCamelCase_ ) _UpperCamelCase = alt_pipe.to(lowerCamelCase_ ) alt_pipe.set_progress_bar_config(disable=lowerCamelCase_ ) _UpperCamelCase = "A painting of a squirrel eating a burger" _UpperCamelCase = torch.manual_seed(0 ) _UpperCamelCase = alt_pipe([prompt] , generator=lowerCamelCase_ , num_inference_steps=2 , output_type="numpy" ) _UpperCamelCase = output.images _UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _UpperCamelCase = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
589
from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image __lowerCAmelCase = ["""text""", """image""", """audio"""] def _lowercase ( a__ : List[str] ) -> List[str]: """simple docstring""" _UpperCamelCase = [] for input_type in input_types: if input_type == "text": inputs.append("Text input" ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" ).resize((5_12, 5_12) ) ) elif input_type == "audio": inputs.append(torch.ones(30_00 ) ) elif isinstance(a__ , a__ ): inputs.append(create_inputs(a__ ) ) else: raise ValueError(f'''Invalid type requested: {input_type}''' ) return inputs def _lowercase ( a__ : List ) -> str: """simple docstring""" _UpperCamelCase = [] for output in outputs: if isinstance(a__ , (str, AgentText) ): output_types.append("text" ) elif isinstance(a__ , (Image.Image, AgentImage) ): output_types.append("image" ) elif isinstance(a__ , (torch.Tensor, AgentAudio) ): output_types.append("audio" ) else: raise ValueError(f'''Invalid output: {output}''' ) return output_types @is_tool_test class lowerCamelCase_ : def lowercase ( self ) -> List[Any]: """simple docstring""" self.assertTrue(hasattr(self.tool , "inputs" ) ) self.assertTrue(hasattr(self.tool , "outputs" ) ) _UpperCamelCase = self.tool.inputs for _input in inputs: if isinstance(_input , lowerCamelCase_ ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) _UpperCamelCase = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def lowercase ( self ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = create_inputs(self.tool.inputs ) _UpperCamelCase = self.tool(*lowerCamelCase_ ) # There is a single output if len(self.tool.outputs ) == 1: _UpperCamelCase = [outputs] self.assertListEqual(output_types(lowerCamelCase_ ) , self.tool.outputs ) def lowercase ( self ) -> str: """simple docstring""" self.assertTrue(hasattr(self.tool , "description" ) ) self.assertTrue(hasattr(self.tool , "default_checkpoint" ) ) self.assertTrue(self.tool.description.startswith("This is a tool that" ) ) def lowercase ( self ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = create_inputs(self.tool.inputs ) _UpperCamelCase = self.tool(*lowerCamelCase_ ) if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): _UpperCamelCase = [outputs] self.assertEqual(len(lowerCamelCase_ ) , len(self.tool.outputs ) ) for output, output_type in zip(lowerCamelCase_ , self.tool.outputs ): _UpperCamelCase = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(lowerCamelCase_ , lowerCamelCase_ ) ) def lowercase ( self ) -> List[Any]: """simple docstring""" _UpperCamelCase = create_inputs(self.tool.inputs ) _UpperCamelCase = [] for _input, input_type in zip(lowerCamelCase_ , self.tool.inputs ): if isinstance(lowerCamelCase_ , lowerCamelCase_ ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error _UpperCamelCase = self.tool(*lowerCamelCase_ ) if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): _UpperCamelCase = [outputs] self.assertEqual(len(lowerCamelCase_ ) , len(self.tool.outputs ) )
589
1
from __future__ import annotations def _lowerCamelCase ( lowerCamelCase_: list[int] ): '''simple docstring''' if not nums: return 0 A : List[str] = nums[0] A : List[str] = 0 for num in nums[1:]: A , A : List[str] = ( max_excluding + num, max(lowerCamelCase_ , lowerCamelCase_ ), ) return max(lowerCamelCase_ , lowerCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
256
from ..utils import DummyObject, requires_backends class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : List[Any] , *snake_case_ : Optional[Any] , **snake_case_ : List[Any] ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Dict , *snake_case_ : Optional[Any] , **snake_case_ : List[str] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : List[Any] , *snake_case_ : Union[str, Any] , **snake_case_ : Dict ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Optional[Any] , *snake_case_ : Tuple , **snake_case_ : str ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Optional[Any] , *snake_case_ : Union[str, Any] , **snake_case_ : Optional[Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Optional[int] , *snake_case_ : Tuple , **snake_case_ : int ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : List[Any] , *snake_case_ : Tuple , **snake_case_ : int ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : List[str] , *snake_case_ : List[str] , **snake_case_ : Optional[Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Optional[Any] , *snake_case_ : str , **snake_case_ : Optional[int] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Any , *snake_case_ : Tuple , **snake_case_ : Dict ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Union[str, Any] , *snake_case_ : List[str] , **snake_case_ : Union[str, Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : List[Any] , *snake_case_ : int , **snake_case_ : Any ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Optional[int] , *snake_case_ : Tuple , **snake_case_ : Union[str, Any] ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Tuple , *snake_case_ : Union[str, Any] , **snake_case_ : List[Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Dict , *snake_case_ : str , **snake_case_ : List[str] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Tuple , *snake_case_ : Union[str, Any] , **snake_case_ : str ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : List[Any] , *snake_case_ : Tuple , **snake_case_ : Optional[int] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : List[Any] , *snake_case_ : Tuple , **snake_case_ : Tuple ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Union[str, Any] , *snake_case_ : str , **snake_case_ : str ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : str , *snake_case_ : Union[str, Any] , **snake_case_ : Union[str, Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Dict , *snake_case_ : int , **snake_case_ : Dict ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : int , *snake_case_ : Any , **snake_case_ : List[str] ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Optional[int] , *snake_case_ : Any , **snake_case_ : Union[str, Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Tuple , *snake_case_ : List[Any] , **snake_case_ : List[str] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Dict , *snake_case_ : int , **snake_case_ : Optional[Any] ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : str , *snake_case_ : int , **snake_case_ : List[Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Optional[int] , *snake_case_ : Optional[Any] , **snake_case_ : int ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : List[Any] , *snake_case_ : Optional[int] , **snake_case_ : Any ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Optional[Any] , *snake_case_ : List[str] , **snake_case_ : Union[str, Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Optional[Any] , *snake_case_ : List[Any] , **snake_case_ : Optional[Any] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Tuple , *snake_case_ : Any , **snake_case_ : Dict ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Dict , *snake_case_ : List[str] , **snake_case_ : Any ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Dict , *snake_case_ : List[Any] , **snake_case_ : Optional[int] ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : Optional[Any] , *snake_case_ : Optional[int] , **snake_case_ : int ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Tuple , *snake_case_ : Union[str, Any] , **snake_case_ : str ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : int , *snake_case_ : Dict , **snake_case_ : Dict ): """simple docstring""" requires_backends(cls , ['''flax'''] ) class _SCREAMING_SNAKE_CASE ( metaclass=snake_case ): lowerCamelCase_ = ['flax'] def __init__( self : str , *snake_case_ : Optional[int] , **snake_case_ : List[Any] ): """simple docstring""" requires_backends(self , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : str , *snake_case_ : Optional[int] , **snake_case_ : Dict ): """simple docstring""" requires_backends(cls , ['''flax'''] ) @classmethod def _UpperCAmelCase ( cls : Any , *snake_case_ : Optional[int] , **snake_case_ : str ): """simple docstring""" requires_backends(cls , ['''flax'''] )
256
1
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def snake_case__ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=1024 , lowerCamelCase_=1024 , lowerCamelCase_=False , **lowerCamelCase_ ): A : List[str] = AutoTokenizer.from_pretrained(lowerCamelCase_ ) A : Optional[int] = SeqaSeqDataset(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , type_path='''train''' , **lowerCamelCase_ ) A : Optional[int] = tok.pad_token_id def get_lens(lowerCamelCase_ ): A : str = tqdm( DataLoader(lowerCamelCase_ , batch_size=512 , num_workers=8 , shuffle=lowerCamelCase_ , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) A : Optional[Any] = [] for batch in dl: A : List[str] = batch['''input_ids'''].ne(lowerCamelCase_ ).sum(1 ).tolist() A : Optional[int] = batch['''labels'''].ne(lowerCamelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCamelCase_ , lowerCamelCase_ ): max_lens.append(max(lowerCamelCase_ , lowerCamelCase_ ) ) else: max_lens.extend(lowerCamelCase_ ) return max_lens A : Any = get_lens(lowerCamelCase_ ) A : Optional[int] = SeqaSeqDataset(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , type_path='''val''' , **lowerCamelCase_ ) A : int = get_lens(lowerCamelCase_ ) pickle_save(lowerCamelCase_ , train_ds.len_file ) pickle_save(lowerCamelCase_ , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
423
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase : List[str] = logging.get_logger(__name__) lowercase : Optional[int] = { "nvidia/segformer-b0-finetuned-ade-512-512": ( "https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json" ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class __lowercase ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ : Dict = '''segformer''' def __init__( self , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=[2, 2, 2, 2] , __UpperCAmelCase=[8, 4, 2, 1] , __UpperCAmelCase=[32, 64, 1_60, 2_56] , __UpperCAmelCase=[7, 3, 3, 3] , __UpperCAmelCase=[4, 2, 2, 2] , __UpperCAmelCase=[1, 2, 5, 8] , __UpperCAmelCase=[4, 4, 4, 4] , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0_2 , __UpperCAmelCase=0.1 , __UpperCAmelCase=1E-6 , __UpperCAmelCase=2_56 , __UpperCAmelCase=2_55 , **__UpperCAmelCase , ) -> Union[str, Any]: super().__init__(**__UpperCAmelCase ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( '''Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be''' ''' removed, as the behaviour will default to that of reshape_last_stage = True.''' , __UpperCAmelCase , ) A : Optional[int] = num_channels A : int = num_encoder_blocks A : Optional[Any] = depths A : List[str] = sr_ratios A : List[Any] = hidden_sizes A : Optional[Any] = patch_sizes A : Any = strides A : Dict = mlp_ratios A : Optional[Any] = num_attention_heads A : int = hidden_act A : Optional[int] = hidden_dropout_prob A : Any = attention_probs_dropout_prob A : Optional[int] = classifier_dropout_prob A : List[Any] = initializer_range A : int = drop_path_rate A : Union[str, Any] = layer_norm_eps A : Union[str, Any] = decoder_hidden_size A : int = kwargs.get('''reshape_last_stage''' , __UpperCAmelCase ) A : str = semantic_loss_ignore_index class __lowercase ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ : List[Any] = version.parse('''1.11''' ) @property def snake_case ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def snake_case ( self ) -> float: return 1E-4 @property def snake_case ( self ) -> int: return 12
423
1
"""simple docstring""" from __future__ import annotations class UpperCAmelCase : def __init__( self : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str ): """simple docstring""" _snake_case = text, pattern _snake_case = len(_SCREAMING_SNAKE_CASE ), len(_SCREAMING_SNAKE_CASE ) def __UpperCAmelCase ( self : int , __lowerCamelCase : Any ): """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : List[str] ): """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def __UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" # searches pattern in text and returns index positions _snake_case = [] for i in range(self.textLen - self.patLen + 1 ): _snake_case = self.mismatch_in_text(_SCREAMING_SNAKE_CASE ) if mismatch_index == -1: positions.append(_SCREAMING_SNAKE_CASE ) else: _snake_case = self.match_in_pattern(self.text[mismatch_index] ) _snake_case = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions snake_case = '''ABAABA''' snake_case = '''AB''' snake_case = BoyerMooreSearch(text, pattern) snake_case = bms.bad_character_heuristic() if len(positions) == 0: print('''No match found''') else: print('''Pattern found in following positions: ''') print(positions)
103
"""simple docstring""" import logging from transformers.configuration_utils import PretrainedConfig UpperCamelCase = logging.getLogger(__name__) class UpperCAmelCase__ ( __lowerCamelCase ): """simple docstring""" lowerCAmelCase__ : str = """masked_bert""" def __init__( self , _SCREAMING_SNAKE_CASE=3_0_5_2_2 , _SCREAMING_SNAKE_CASE=7_6_8 , _SCREAMING_SNAKE_CASE=1_2 , _SCREAMING_SNAKE_CASE=1_2 , _SCREAMING_SNAKE_CASE=3_0_7_2 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=5_1_2 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.0_2 , _SCREAMING_SNAKE_CASE=1E-12 , _SCREAMING_SNAKE_CASE=0 , _SCREAMING_SNAKE_CASE="topK" , _SCREAMING_SNAKE_CASE="constant" , _SCREAMING_SNAKE_CASE=0.0 , **_SCREAMING_SNAKE_CASE , ) -> List[Any]: super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) a_ : Union[str, Any] = vocab_size a_ : Optional[int] = hidden_size a_ : Any = num_hidden_layers a_ : Tuple = num_attention_heads a_ : Dict = hidden_act a_ : Dict = intermediate_size a_ : Optional[int] = hidden_dropout_prob a_ : int = attention_probs_dropout_prob a_ : int = max_position_embeddings a_ : Union[str, Any] = type_vocab_size a_ : Optional[Any] = initializer_range a_ : Union[str, Any] = layer_norm_eps a_ : Tuple = pruning_method a_ : Tuple = mask_init a_ : Dict = mask_scale
473
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase : Dict = logging.get_logger(__name__) _lowerCamelCase : Optional[int] = { "facebook/s2t-small-librispeech-asr": ( "https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech_to_text } class __snake_case (_a ): lowerCAmelCase__ = "speech_to_text" lowerCAmelCase__ = ["past_key_values"] lowerCAmelCase__ = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : Optional[int] , _UpperCAmelCase : Optional[int]=1_0000 , _UpperCAmelCase : List[str]=12 , _UpperCAmelCase : Optional[Any]=2048 , _UpperCAmelCase : Tuple=4 , _UpperCAmelCase : Optional[Any]=6 , _UpperCAmelCase : Optional[int]=2048 , _UpperCAmelCase : Any=4 , _UpperCAmelCase : Dict=0.0 , _UpperCAmelCase : List[Any]=0.0 , _UpperCAmelCase : Dict=True , _UpperCAmelCase : str=True , _UpperCAmelCase : Any="relu" , _UpperCAmelCase : str=256 , _UpperCAmelCase : Tuple=0.1 , _UpperCAmelCase : Tuple=0.0 , _UpperCAmelCase : List[Any]=0.0 , _UpperCAmelCase : Dict=0.02 , _UpperCAmelCase : Union[str, Any]=2 , _UpperCAmelCase : Dict=True , _UpperCAmelCase : Any=1 , _UpperCAmelCase : List[str]=0 , _UpperCAmelCase : Dict=2 , _UpperCAmelCase : List[Any]=6000 , _UpperCAmelCase : Optional[Any]=1024 , _UpperCAmelCase : List[Any]=2 , _UpperCAmelCase : str=(5, 5) , _UpperCAmelCase : Optional[Any]=1024 , _UpperCAmelCase : List[Any]=80 , _UpperCAmelCase : Any=1 , **_UpperCAmelCase : Any , ) -> Any: '''simple docstring''' _lowerCAmelCase : Any = vocab_size _lowerCAmelCase : Optional[Any] = d_model _lowerCAmelCase : Optional[int] = encoder_ffn_dim _lowerCAmelCase : List[str] = encoder_layers _lowerCAmelCase : Tuple = encoder_attention_heads _lowerCAmelCase : Any = decoder_ffn_dim _lowerCAmelCase : Optional[Any] = decoder_layers _lowerCAmelCase : Union[str, Any] = decoder_attention_heads _lowerCAmelCase : List[str] = dropout _lowerCAmelCase : List[Any] = attention_dropout _lowerCAmelCase : List[Any] = activation_dropout _lowerCAmelCase : Optional[Any] = activation_function _lowerCAmelCase : Tuple = init_std _lowerCAmelCase : str = encoder_layerdrop _lowerCAmelCase : Tuple = decoder_layerdrop _lowerCAmelCase : str = use_cache _lowerCAmelCase : Optional[Any] = encoder_layers _lowerCAmelCase : Any = scale_embedding # scale factor will be sqrt(d_model) if True _lowerCAmelCase : Dict = max_source_positions _lowerCAmelCase : Optional[int] = max_target_positions _lowerCAmelCase : Tuple = num_conv_layers _lowerCAmelCase : Any = list(_UpperCAmelCase ) _lowerCAmelCase : int = conv_channels _lowerCAmelCase : Optional[int] = input_feat_per_channel _lowerCAmelCase : Optional[int] = input_channels if len(self.conv_kernel_sizes ) != self.num_conv_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.conv_kernel_sizes)` == `config.num_conv_layers` """ f"but is `len(config.conv_kernel_sizes) = {len(self.conv_kernel_sizes )}`, " f"`config.num_conv_layers = {self.num_conv_layers}`." ) super().__init__( pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , is_encoder_decoder=_UpperCAmelCase , decoder_start_token_id=_UpperCAmelCase , **_UpperCAmelCase , )
196
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _lowerCamelCase : str = get_logger(__name__) class __snake_case : def __init__( self : Optional[int] , _UpperCAmelCase : Optional[str] = None ) -> List[str]: '''simple docstring''' _lowerCAmelCase : Optional[Any] = ( os.path.join(_UpperCAmelCase , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) _lowerCAmelCase : Optional[int] = Extractor def SCREAMING_SNAKE_CASE ( self : Tuple , _UpperCAmelCase : str ) -> str: '''simple docstring''' from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" _lowerCAmelCase : List[Any] = os.path.abspath(_UpperCAmelCase ) return os.path.join(self.extract_dir , hash_url_to_filename(_UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , _UpperCAmelCase : str , _UpperCAmelCase : bool ) -> bool: '''simple docstring''' return force_extract or ( not os.path.isfile(_UpperCAmelCase ) and not (os.path.isdir(_UpperCAmelCase ) and os.listdir(_UpperCAmelCase )) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , _UpperCAmelCase : str , _UpperCAmelCase : bool = False ) -> str: '''simple docstring''' _lowerCAmelCase : str = self.extractor.infer_extractor_format(_UpperCAmelCase ) if not extractor_format: return input_path _lowerCAmelCase : int = self._get_output_path(_UpperCAmelCase ) if self._do_extract(_UpperCAmelCase , _UpperCAmelCase ): self.extractor.extract(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return output_path class __snake_case (_a ): @classmethod @abstractmethod def SCREAMING_SNAKE_CASE ( cls : Tuple , _UpperCAmelCase : Union[Path, str] , **_UpperCAmelCase : Optional[int] ) -> bool: '''simple docstring''' ... @staticmethod @abstractmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' ... class __snake_case (_a , _a ): lowerCAmelCase__ = [] @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : int ) -> Dict: '''simple docstring''' with open(_UpperCAmelCase , """rb""" ) as f: return f.read(_UpperCAmelCase ) @classmethod def SCREAMING_SNAKE_CASE ( cls : Any , _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : bytes = b"" ) -> bool: '''simple docstring''' if not magic_number: _lowerCAmelCase : Optional[int] = max(len(_UpperCAmelCase ) for cls_magic_number in cls.magic_numbers ) try: _lowerCAmelCase : Union[str, Any] = cls.read_magic_number(_UpperCAmelCase , _UpperCAmelCase ) except OSError: return False return any(magic_number.startswith(_UpperCAmelCase ) for cls_magic_number in cls.magic_numbers ) class __snake_case (_a ): @classmethod def SCREAMING_SNAKE_CASE ( cls : str , _UpperCAmelCase : Union[Path, str] , **_UpperCAmelCase : Dict ) -> bool: '''simple docstring''' return tarfile.is_tarfile(_UpperCAmelCase ) @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple ) -> Tuple: '''simple docstring''' def resolved(_UpperCAmelCase : str ) -> str: return os.path.realpath(os.path.abspath(_UpperCAmelCase ) ) def badpath(_UpperCAmelCase : str , _UpperCAmelCase : str ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ).startswith(_UpperCAmelCase ) def badlink(_UpperCAmelCase : Optional[int] , _UpperCAmelCase : str ) -> bool: # Links are interpreted relative to the directory containing the link _lowerCAmelCase : Tuple = resolved(os.path.join(_UpperCAmelCase , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=_UpperCAmelCase ) _lowerCAmelCase : List[Any] = resolved(_UpperCAmelCase ) for finfo in members: if badpath(finfo.name , _UpperCAmelCase ): logger.error(f"Extraction of {finfo.name} is blocked (illegal path)" ) elif finfo.issym() and badlink(_UpperCAmelCase , _UpperCAmelCase ): logger.error(f"Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}" ) elif finfo.islnk() and badlink(_UpperCAmelCase , _UpperCAmelCase ): logger.error(f"Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}" ) else: yield finfo @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) _lowerCAmelCase : str = tarfile.open(_UpperCAmelCase ) tar_file.extractall(_UpperCAmelCase , members=TarExtractor.safemembers(_UpperCAmelCase , _UpperCAmelCase ) ) tar_file.close() class __snake_case (_a ): lowerCAmelCase__ = [b"\x1F\x8B"] @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' with gzip.open(_UpperCAmelCase , """rb""" ) as gzip_file: with open(_UpperCAmelCase , """wb""" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class __snake_case (_a ): lowerCAmelCase__ = [ b"PK\x03\x04", b"PK\x05\x06", # empty archive b"PK\x07\x08", # spanned archive ] @classmethod def SCREAMING_SNAKE_CASE ( cls : Dict , _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : bytes = b"" ) -> bool: '''simple docstring''' if super().is_extractable(_UpperCAmelCase , magic_number=_UpperCAmelCase ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(_UpperCAmelCase , """rb""" ) as fp: _lowerCAmelCase : Union[str, Any] = _EndRecData(_UpperCAmelCase ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: _lowerCAmelCase : List[Any] = fp.read(_UpperCAmelCase ) # CD is where we expect it to be if len(_UpperCAmelCase ) == sizeCentralDir: _lowerCAmelCase : int = struct.unpack(_UpperCAmelCase , _UpperCAmelCase ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) with zipfile.ZipFile(_UpperCAmelCase , """r""" ) as zip_file: zip_file.extractall(_UpperCAmelCase ) zip_file.close() class __snake_case (_a ): lowerCAmelCase__ = [b"\xFD\x37\x7A\x58\x5A\x00"] @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' with lzma.open(_UpperCAmelCase ) as compressed_file: with open(_UpperCAmelCase , """wb""" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class __snake_case (_a ): lowerCAmelCase__ = [b"Rar!\x1a\x07\x00", b"Rar!\x1a\x07\x01\x00"] # RAR_ID # RAR5_ID @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' if not config.RARFILE_AVAILABLE: raise ImportError("""Please pip install rarfile""" ) import rarfile os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) _lowerCAmelCase : str = rarfile.RarFile(_UpperCAmelCase ) rf.extractall(_UpperCAmelCase ) rf.close() class __snake_case (_a ): lowerCAmelCase__ = [b"\x28\xb5\x2F\xFD"] @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' if not config.ZSTANDARD_AVAILABLE: raise ImportError("""Please pip install zstandard""" ) import zstandard as zstd _lowerCAmelCase : Any = zstd.ZstdDecompressor() with open(_UpperCAmelCase , """rb""" ) as ifh, open(_UpperCAmelCase , """wb""" ) as ofh: dctx.copy_stream(_UpperCAmelCase , _UpperCAmelCase ) class __snake_case (_a ): lowerCAmelCase__ = [b"\x42\x5A\x68"] @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' with bza.open(_UpperCAmelCase , """rb""" ) as compressed_file: with open(_UpperCAmelCase , """wb""" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class __snake_case (_a ): lowerCAmelCase__ = [b"\x37\x7A\xBC\xAF\x27\x1C"] @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' if not config.PY7ZR_AVAILABLE: raise ImportError("""Please pip install py7zr""" ) import pyazr os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) with pyazr.SevenZipFile(_UpperCAmelCase , """r""" ) as archive: archive.extractall(_UpperCAmelCase ) class __snake_case (_a ): lowerCAmelCase__ = [b"\x04\x22\x4D\x18"] @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] ) -> None: '''simple docstring''' if not config.LZ4_AVAILABLE: raise ImportError("""Please pip install lz4""" ) import lza.frame with lza.frame.open(_UpperCAmelCase , """rb""" ) as compressed_file: with open(_UpperCAmelCase , """wb""" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class __snake_case : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) lowerCAmelCase__ = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def SCREAMING_SNAKE_CASE ( cls : str ) -> Optional[int]: '''simple docstring''' return max( len(_UpperCAmelCase ) for extractor in cls.extractors.values() if issubclass(_UpperCAmelCase , _UpperCAmelCase ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def SCREAMING_SNAKE_CASE ( _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : int ) -> Any: '''simple docstring''' try: return MagicNumberBaseExtractor.read_magic_number(_UpperCAmelCase , magic_number_length=_UpperCAmelCase ) except OSError: return b"" @classmethod def SCREAMING_SNAKE_CASE ( cls : int , _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : bool = False ) -> bool: '''simple docstring''' warnings.warn( """Method 'is_extractable' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'infer_extractor_format' instead.""" , category=_UpperCAmelCase , ) _lowerCAmelCase : str = cls.infer_extractor_format(_UpperCAmelCase ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def SCREAMING_SNAKE_CASE ( cls : Optional[Any] , _UpperCAmelCase : Union[Path, str] ) -> str: # <Added version="2.4.0"/> '''simple docstring''' _lowerCAmelCase : str = cls._get_magic_number_max_length() _lowerCAmelCase : Dict = cls._read_magic_number(_UpperCAmelCase , _UpperCAmelCase ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(_UpperCAmelCase , magic_number=_UpperCAmelCase ): return extractor_format @classmethod def SCREAMING_SNAKE_CASE ( cls : Optional[Any] , _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Union[Path, str] , _UpperCAmelCase : Optional[str] = None , _UpperCAmelCase : Optional[BaseExtractor] = "deprecated" , ) -> None: '''simple docstring''' os.makedirs(os.path.dirname(_UpperCAmelCase ) , exist_ok=_UpperCAmelCase ) # Prevent parallel extractions _lowerCAmelCase : Tuple = str(Path(_UpperCAmelCase ).with_suffix(""".lock""" ) ) with FileLock(_UpperCAmelCase ): shutil.rmtree(_UpperCAmelCase , ignore_errors=_UpperCAmelCase ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(_UpperCAmelCase , _UpperCAmelCase ): # passed as positional arg warnings.warn( """Parameter 'extractor' was deprecated in version 2.4.0 and will be removed in 3.0.0. """ """Use 'extractor_format' instead.""" , category=_UpperCAmelCase , ) _lowerCAmelCase : List[str] = extractor if extractor != """deprecated""" else extractor_format else: _lowerCAmelCase : List[Any] = cls.extractors[extractor_format] return extractor.extract(_UpperCAmelCase , _UpperCAmelCase ) else: warnings.warn( """Parameter 'extractor_format' was made required in version 2.4.0 and not passing it will raise an """ """exception in 3.0.0.""" , category=_UpperCAmelCase , ) for extractor in cls.extractors.values(): if extractor.is_extractable(_UpperCAmelCase ): return extractor.extract(_UpperCAmelCase , _UpperCAmelCase )
196
1
from typing import Union import fire import torch from tqdm import tqdm def a__ ( snake_case__ : str , snake_case__ : str = "cpu" , snake_case__ : Union[str, None] = None ): _UpperCAmelCase : Optional[Any] = torch.load(snake_case__ , map_location=snake_case__ ) for k, v in tqdm(state_dict.items() ): if not isinstance(snake_case__ , torch.Tensor ): raise TypeError("""FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin""" ) _UpperCAmelCase : int = v.half() if save_path is None: # overwrite src_path _UpperCAmelCase : Any = src_path torch.save(snake_case__ , snake_case__ ) if __name__ == "__main__": fire.Fire(convert)
643
import collections import importlib.util import os import re from pathlib import Path SCREAMING_SNAKE_CASE__ : List[Any] = 'src/transformers' # Matches is_xxx_available() SCREAMING_SNAKE_CASE__ : List[Any] = re.compile(R'is\_([a-z_]*)_available()') # Catches a one-line _import_struct = {xxx} SCREAMING_SNAKE_CASE__ : str = re.compile(R'^_import_structure\s+=\s+\{([^\}]+)\}') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] SCREAMING_SNAKE_CASE__ : Tuple = re.compile(R'\s+"\S*":\s+\[([^\]]*)\]') # Catches a line if not is_foo_available SCREAMING_SNAKE_CASE__ : Dict = re.compile(R'^\s*if\s+not\s+is\_[a-z_]*\_available\(\)') # Catches a line _import_struct["bla"].append("foo") SCREAMING_SNAKE_CASE__ : Union[str, Any] = re.compile(R'^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] SCREAMING_SNAKE_CASE__ : Optional[Any] = re.compile(R'^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]') # Catches a line with an object between quotes and a comma: "MyModel", SCREAMING_SNAKE_CASE__ : Optional[int] = re.compile('^\s+"([^"]+)",') # Catches a line with objects between brackets only: ["foo", "bar"], SCREAMING_SNAKE_CASE__ : Union[str, Any] = re.compile('^\s+\[([^\]]+)\]') # Catches a line with from foo import bar, bla, boo SCREAMING_SNAKE_CASE__ : List[Any] = re.compile(R'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') # Catches a line with try: SCREAMING_SNAKE_CASE__ : Optional[Any] = re.compile(R'^\s*try:') # Catches a line with else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = re.compile(R'^\s*else:') def a__ ( snake_case__ : Union[str, Any] ): if _re_test_backend.search(snake_case__ ) is None: return None _UpperCAmelCase : str = [b[0] for b in _re_backend.findall(snake_case__ )] backends.sort() return "_and_".join(snake_case__ ) def a__ ( snake_case__ : Optional[int] ): with open(snake_case__ , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: _UpperCAmelCase : Optional[Any] = f.readlines() _UpperCAmelCase : int = 0 while line_index < len(snake_case__ ) and not lines[line_index].startswith("""_import_structure = {""" ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(snake_case__ ): return None # First grab the objects without a specific backend in _import_structure _UpperCAmelCase : int = [] while not lines[line_index].startswith("""if TYPE_CHECKING""" ) and find_backend(lines[line_index] ) is None: _UpperCAmelCase : int = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(snake_case__ ): _UpperCAmelCase : Optional[Any] = _re_one_line_import_struct.search(snake_case__ ).groups()[0] _UpperCAmelCase : str = re.findall("""\[([^\]]+)\]""" , snake_case__ ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(""", """ )] ) line_index += 1 continue _UpperCAmelCase : int = _re_import_struct_key_value.search(snake_case__ ) if single_line_import_search is not None: _UpperCAmelCase : Dict = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(""", """ ) if len(snake_case__ ) > 0] objects.extend(snake_case__ ) elif line.startswith(""" """ * 8 + """\"""" ): objects.append(line[9:-3] ) line_index += 1 _UpperCAmelCase : str = {"""none""": objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith("""if TYPE_CHECKING""" ): # If the line is an if not is_backend_available, we grab all objects associated. _UpperCAmelCase : Tuple = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _UpperCAmelCase : List[Any] = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _UpperCAmelCase : List[Any] = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(""" """ * 4 ): _UpperCAmelCase : Optional[int] = lines[line_index] if _re_import_struct_add_one.search(snake_case__ ) is not None: objects.append(_re_import_struct_add_one.search(snake_case__ ).groups()[0] ) elif _re_import_struct_add_many.search(snake_case__ ) is not None: _UpperCAmelCase : str = _re_import_struct_add_many.search(snake_case__ ).groups()[0].split(""", """ ) _UpperCAmelCase : Optional[Any] = [obj[1:-1] for obj in imports if len(snake_case__ ) > 0] objects.extend(snake_case__ ) elif _re_between_brackets.search(snake_case__ ) is not None: _UpperCAmelCase : str = _re_between_brackets.search(snake_case__ ).groups()[0].split(""", """ ) _UpperCAmelCase : List[str] = [obj[1:-1] for obj in imports if len(snake_case__ ) > 0] objects.extend(snake_case__ ) elif _re_quote_object.search(snake_case__ ) is not None: objects.append(_re_quote_object.search(snake_case__ ).groups()[0] ) elif line.startswith(""" """ * 8 + """\"""" ): objects.append(line[9:-3] ) elif line.startswith(""" """ * 12 + """\"""" ): objects.append(line[13:-3] ) line_index += 1 _UpperCAmelCase : str = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend _UpperCAmelCase : Optional[Any] = [] while ( line_index < len(snake_case__ ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith("""else""" ) ): _UpperCAmelCase : Union[str, Any] = lines[line_index] _UpperCAmelCase : str = _re_import.search(snake_case__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(""", """ ) ) elif line.startswith(""" """ * 8 ): objects.append(line[8:-2] ) line_index += 1 _UpperCAmelCase : int = {"""none""": objects} # Let's continue with backend-specific objects while line_index < len(snake_case__ ): # If the line is an if is_backend_available, we grab all objects associated. _UpperCAmelCase : str = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _UpperCAmelCase : Optional[int] = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _UpperCAmelCase : Dict = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(""" """ * 8 ): _UpperCAmelCase : Union[str, Any] = lines[line_index] _UpperCAmelCase : Any = _re_import.search(snake_case__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(""", """ ) ) elif line.startswith(""" """ * 12 ): objects.append(line[12:-2] ) line_index += 1 _UpperCAmelCase : str = objects else: line_index += 1 return import_dict_objects, type_hint_objects def a__ ( snake_case__ : Any , snake_case__ : Optional[int] ): def find_duplicates(snake_case__ : Dict ): return [k for k, v in collections.Counter(snake_case__ ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] _UpperCAmelCase : int = [] for key in import_dict_objects.keys(): _UpperCAmelCase : Any = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(f'''Duplicate _import_structure definitions for: {duplicate_imports}''' ) _UpperCAmelCase : Optional[int] = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(f'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): _UpperCAmelCase : Optional[int] = """base imports""" if key == """none""" else f'''{key} backend''' errors.append(f'''Differences for {name}:''' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f''' {a} in TYPE_HINT but not in _import_structure.''' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f''' {a} in _import_structure but not in TYPE_HINT.''' ) return errors def a__ ( ): _UpperCAmelCase : Any = [] for root, _, files in os.walk(snake_case__ ): if "__init__.py" in files: _UpperCAmelCase : Optional[Any] = os.path.join(snake_case__ , """__init__.py""" ) _UpperCAmelCase : List[str] = parse_init(snake_case__ ) if objects is not None: _UpperCAmelCase : int = analyze_results(*snake_case__ ) if len(snake_case__ ) > 0: _UpperCAmelCase : Union[str, Any] = f'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append("""\n""".join(snake_case__ ) ) if len(snake_case__ ) > 0: raise ValueError("""\n\n""".join(snake_case__ ) ) def a__ ( ): _UpperCAmelCase : Tuple = [] for path, directories, files in os.walk(snake_case__ ): for folder in directories: # Ignore private modules if folder.startswith("""_""" ): directories.remove(snake_case__ ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(snake_case__ ) / folder).glob("""*.py""" ) ) ) == 0: continue _UpperCAmelCase : Dict = str((Path(snake_case__ ) / folder).relative_to(snake_case__ ) ) _UpperCAmelCase : Union[str, Any] = short_path.replace(os.path.sep , """.""" ) submodules.append(snake_case__ ) for fname in files: if fname == "__init__.py": continue _UpperCAmelCase : Optional[Any] = str((Path(snake_case__ ) / fname).relative_to(snake_case__ ) ) _UpperCAmelCase : Optional[Any] = short_path.replace(""".py""" , """""" ).replace(os.path.sep , """.""" ) if len(submodule.split(""".""" ) ) == 1: submodules.append(snake_case__ ) return submodules SCREAMING_SNAKE_CASE__ : List[Any] = [ 'convert_pytorch_checkpoint_to_tf2', 'modeling_flax_pytorch_utils', ] def a__ ( ): # This is to make sure the transformers module imported is the one in the repo. _UpperCAmelCase : int = importlib.util.spec_from_file_location( """transformers""" , os.path.join(snake_case__ , """__init__.py""" ) , submodule_search_locations=[PATH_TO_TRANSFORMERS] , ) _UpperCAmelCase : Optional[int] = spec.loader.load_module() _UpperCAmelCase : int = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(snake_case__ ) > 0: _UpperCAmelCase : Dict = """\n""".join(f'''- {module}''' for module in module_not_registered ) raise ValueError( """The following submodules are not properly registered in the main init of Transformers:\n""" f'''{list_of_modules}\n''' """Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.""" ) if __name__ == "__main__": check_all_inits() check_submodules()
643
1
lowercase_ : Optional[int] = '0.18.2' from .configuration_utils import ConfigMixin from .utils import ( OptionalDependencyNotAvailable, is_flax_available, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_note_seq_available, is_onnx_available, is_scipy_available, is_torch_available, is_torchsde_available, is_transformers_available, is_transformers_version, is_unidecode_available, logging, ) try: if not is_onnx_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_onnx_objects import * # noqa F403 else: from .pipelines import OnnxRuntimeModel try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_pt_objects import * # noqa F403 else: from .models import ( AutoencoderKL, ControlNetModel, ModelMixin, PriorTransformer, TaFilmDecoder, TransformeraDModel, UNetaDModel, UNetaDConditionModel, UNetaDModel, UNetaDConditionModel, VQModel, ) from .optimization import ( get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, get_scheduler, ) from .pipelines import ( AudioPipelineOutput, ConsistencyModelPipeline, DanceDiffusionPipeline, DDIMPipeline, DDPMPipeline, DiffusionPipeline, DiTPipeline, ImagePipelineOutput, KarrasVePipeline, LDMPipeline, LDMSuperResolutionPipeline, PNDMPipeline, RePaintPipeline, ScoreSdeVePipeline, ) from .schedulers import ( CMStochasticIterativeScheduler, DDIMInverseScheduler, DDIMParallelScheduler, DDIMScheduler, DDPMParallelScheduler, DDPMScheduler, DEISMultistepScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, HeunDiscreteScheduler, IPNDMScheduler, KarrasVeScheduler, KDPMaAncestralDiscreteScheduler, KDPMaDiscreteScheduler, PNDMScheduler, RePaintScheduler, SchedulerMixin, ScoreSdeVeScheduler, UnCLIPScheduler, UniPCMultistepScheduler, VQDiffusionScheduler, ) from .training_utils import EMAModel try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .schedulers import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .schedulers import DPMSolverSDEScheduler try: if not (is_torch_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipelines import ( AltDiffusionImgaImgPipeline, AltDiffusionPipeline, AudioLDMPipeline, CycleDiffusionPipeline, IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ImageTextPipelineOutput, KandinskyImgaImgPipeline, KandinskyInpaintPipeline, KandinskyPipeline, KandinskyPriorPipeline, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaControlnetPipeline, KandinskyVaaImgaImgPipeline, KandinskyVaaInpaintPipeline, KandinskyVaaPipeline, KandinskyVaaPriorEmbaEmbPipeline, KandinskyVaaPriorPipeline, LDMTextToImagePipeline, PaintByExamplePipeline, SemanticStableDiffusionPipeline, ShapEImgaImgPipeline, ShapEPipeline, StableDiffusionAttendAndExcitePipeline, StableDiffusionControlNetImgaImgPipeline, StableDiffusionControlNetInpaintPipeline, StableDiffusionControlNetPipeline, StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionImageVariationPipeline, StableDiffusionImgaImgPipeline, StableDiffusionInpaintPipeline, StableDiffusionInpaintPipelineLegacy, StableDiffusionInstructPixaPixPipeline, StableDiffusionLatentUpscalePipeline, StableDiffusionLDMaDPipeline, StableDiffusionModelEditingPipeline, StableDiffusionPanoramaPipeline, StableDiffusionParadigmsPipeline, StableDiffusionPipeline, StableDiffusionPipelineSafe, StableDiffusionPixaPixZeroPipeline, StableDiffusionSAGPipeline, StableDiffusionUpscalePipeline, StableUnCLIPImgaImgPipeline, StableUnCLIPPipeline, TextToVideoSDPipeline, TextToVideoZeroPipeline, UnCLIPImageVariationPipeline, UnCLIPPipeline, UniDiffuserModel, UniDiffuserPipeline, UniDiffuserTextDecoder, VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, VideoToVideoSDPipeline, VQDiffusionPipeline, ) try: if not (is_torch_available() and is_transformers_available() and is_invisible_watermark_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_invisible_watermark_objects import * # noqa F403 else: from .pipelines import StableDiffusionXLImgaImgPipeline, StableDiffusionXLPipeline try: if not (is_torch_available() and is_transformers_available() and is_k_diffusion_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipelines import StableDiffusionKDiffusionPipeline try: if not (is_torch_available() and is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_onnx_objects import * # noqa F403 else: from .pipelines import ( OnnxStableDiffusionImgaImgPipeline, OnnxStableDiffusionInpaintPipeline, OnnxStableDiffusionInpaintPipelineLegacy, OnnxStableDiffusionPipeline, OnnxStableDiffusionUpscalePipeline, StableDiffusionOnnxPipeline, ) try: if not (is_torch_available() and is_librosa_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_librosa_objects import * # noqa F403 else: from .pipelines import AudioDiffusionPipeline, Mel try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .pipelines import SpectrogramDiffusionPipeline try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_objects import * # noqa F403 else: from .models.controlnet_flax import FlaxControlNetModel from .models.modeling_flax_utils import FlaxModelMixin from .models.unet_ad_condition_flax import FlaxUNetaDConditionModel from .models.vae_flax import FlaxAutoencoderKL from .pipelines import FlaxDiffusionPipeline from .schedulers import ( FlaxDDIMScheduler, FlaxDDPMScheduler, FlaxDPMSolverMultistepScheduler, FlaxKarrasVeScheduler, FlaxLMSDiscreteScheduler, FlaxPNDMScheduler, FlaxSchedulerMixin, FlaxScoreSdeVeScheduler, ) try: if not (is_flax_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_and_transformers_objects import * # noqa F403 else: from .pipelines import ( FlaxStableDiffusionControlNetPipeline, FlaxStableDiffusionImgaImgPipeline, FlaxStableDiffusionInpaintPipeline, FlaxStableDiffusionPipeline, ) try: if not (is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_note_seq_objects import * # noqa F403 else: from .pipelines import MidiProcessor
107
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ : Dict = { 'configuration_autoformer': [ 'AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'AutoformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ : str = [ 'AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'AutoformerForPrediction', 'AutoformerModel', 'AutoformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys lowercase_ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
107
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: lowercase = None lowercase = logging.get_logger(__name__) lowercase = {'vocab_file': 'sentencepiece.model', 'tokenizer_file': 'tokenizer.json'} lowercase = { 'vocab_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/sentencepiece.model', }, 'tokenizer_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/tokenizer.json', }, } lowercase = { 'google/rembert': 2_56, } lowercase = '▁' class lowercase__ ( _lowerCamelCase ): '''simple docstring''' _UpperCAmelCase = VOCAB_FILES_NAMES _UpperCAmelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase = RemBertTokenizer def __init__( self , snake_case=None , snake_case=None , snake_case=True , snake_case=True , snake_case=False , snake_case="[CLS]" , snake_case="[SEP]" , snake_case="<unk>" , snake_case="[SEP]" , snake_case="<pad>" , snake_case="[CLS]" , snake_case="[MASK]" , **snake_case , ) -> Union[str, Any]: _UpperCAmelCase = AddedToken(_SCREAMING_SNAKE_CASE , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else mask_token super().__init__( _SCREAMING_SNAKE_CASE , tokenizer_file=_SCREAMING_SNAKE_CASE , do_lower_case=_SCREAMING_SNAKE_CASE , remove_space=_SCREAMING_SNAKE_CASE , keep_accents=_SCREAMING_SNAKE_CASE , bos_token=_SCREAMING_SNAKE_CASE , eos_token=_SCREAMING_SNAKE_CASE , unk_token=_SCREAMING_SNAKE_CASE , sep_token=_SCREAMING_SNAKE_CASE , pad_token=_SCREAMING_SNAKE_CASE , cls_token=_SCREAMING_SNAKE_CASE , mask_token=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = do_lower_case _UpperCAmelCase = remove_space _UpperCAmelCase = keep_accents _UpperCAmelCase = vocab_file _UpperCAmelCase = False if not self.vocab_file else True def lowerCamelCase_ ( self , snake_case , snake_case = None ) -> List[int]: _UpperCAmelCase = [self.sep_token_id] _UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase_ ( self , snake_case , snake_case = None , snake_case = False ) -> List[int]: if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(_SCREAMING_SNAKE_CASE )) + [1] + ([0] * len(_SCREAMING_SNAKE_CASE )) + [1] return [1] + ([0] * len(_SCREAMING_SNAKE_CASE )) + [1] def lowerCamelCase_ ( self , snake_case , snake_case = None ) -> List[int]: _UpperCAmelCase = [self.sep_token_id] _UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase_ ( self , snake_case , snake_case = None ) -> Tuple[str]: if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error('Vocabulary path ({}) should be a directory'.format(_SCREAMING_SNAKE_CASE ) ) return _UpperCAmelCase = os.path.join( _SCREAMING_SNAKE_CASE , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ): copyfile(self.vocab_file , _SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
573
"""simple docstring""" import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __init__( self : Union[str, Any] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Union[str, Any]=7 , _SCREAMING_SNAKE_CASE : Union[str, Any]=3 , _SCREAMING_SNAKE_CASE : Optional[int]=18 , _SCREAMING_SNAKE_CASE : List[Any]=30 , _SCREAMING_SNAKE_CASE : Optional[int]=400 , _SCREAMING_SNAKE_CASE : Dict=True , _SCREAMING_SNAKE_CASE : str=None , _SCREAMING_SNAKE_CASE : Any=True , _SCREAMING_SNAKE_CASE : Union[str, Any]=None , _SCREAMING_SNAKE_CASE : Dict=True , _SCREAMING_SNAKE_CASE : List[str]=[0.5, 0.5, 0.5] , _SCREAMING_SNAKE_CASE : Optional[int]=[0.5, 0.5, 0.5] , _SCREAMING_SNAKE_CASE : List[Any]=False , ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = size if size is not None else {'height': 20, 'width': 20} SCREAMING_SNAKE_CASE : int = crop_size if crop_size is not None else {'height': 18, 'width': 18} SCREAMING_SNAKE_CASE : Optional[int] = parent SCREAMING_SNAKE_CASE : Dict = batch_size SCREAMING_SNAKE_CASE : Optional[Any] = num_channels SCREAMING_SNAKE_CASE : Tuple = image_size SCREAMING_SNAKE_CASE : Optional[Any] = min_resolution SCREAMING_SNAKE_CASE : List[str] = max_resolution SCREAMING_SNAKE_CASE : Tuple = do_resize SCREAMING_SNAKE_CASE : str = size SCREAMING_SNAKE_CASE : Dict = do_center_crop SCREAMING_SNAKE_CASE : List[Any] = crop_size SCREAMING_SNAKE_CASE : Optional[int] = do_normalize SCREAMING_SNAKE_CASE : Optional[Any] = image_mean SCREAMING_SNAKE_CASE : Dict = image_std SCREAMING_SNAKE_CASE : Any = do_reduce_labels def _lowerCAmelCase ( self : Dict ) -> str: """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def __snake_case ( ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) SCREAMING_SNAKE_CASE : Any = Image.open(dataset[0]['file'] ) SCREAMING_SNAKE_CASE : int = Image.open(dataset[1]['file'] ) return image, map def __snake_case ( ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE : int = load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) SCREAMING_SNAKE_CASE : Optional[int] = Image.open(ds[0]['file'] ) SCREAMING_SNAKE_CASE : Optional[int] = Image.open(ds[1]['file'] ) SCREAMING_SNAKE_CASE : Tuple = Image.open(ds[2]['file'] ) SCREAMING_SNAKE_CASE : Tuple = Image.open(ds[3]['file'] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class lowerCAmelCase__ ( _lowerCamelCase , unittest.TestCase ): '''simple docstring''' _SCREAMING_SNAKE_CASE : Tuple = BeitImageProcessor if is_vision_available() else None def _lowerCAmelCase ( self : Optional[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = BeitImageProcessingTester(self ) @property def _lowerCAmelCase ( self : Optional[int] ) -> int: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowerCAmelCase ( self : int ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , 'do_resize' ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , 'size' ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , 'do_center_crop' ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , 'center_crop' ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , 'do_normalize' ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , 'image_mean' ) ) self.assertTrue(hasattr(_SCREAMING_SNAKE_CASE , 'image_std' ) ) def _lowerCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE : int = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 20, 'width': 20} ) self.assertEqual(image_processor.crop_size , {'height': 18, 'width': 18} ) self.assertEqual(image_processor.do_reduce_labels , _SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : List[Any] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=_SCREAMING_SNAKE_CASE ) self.assertEqual(image_processor.size , {'height': 42, 'width': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) self.assertEqual(image_processor.do_reduce_labels , _SCREAMING_SNAKE_CASE ) def _lowerCAmelCase ( self : str ) -> Dict: """simple docstring""" pass def _lowerCAmelCase ( self : Any ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE : str = image_processing(_SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def _lowerCAmelCase ( self : Union[str, Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE , numpify=_SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE : Dict = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE : str = image_processing(_SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def _lowerCAmelCase ( self : str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE , torchify=_SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE : List[str] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE : str = image_processing(_SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def _lowerCAmelCase ( self : int ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=_SCREAMING_SNAKE_CASE , torchify=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : Tuple = [] for image in image_inputs: self.assertIsInstance(_SCREAMING_SNAKE_CASE , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input SCREAMING_SNAKE_CASE : int = image_processing(image_inputs[0] , maps[0] , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( 1, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) # Test batched SCREAMING_SNAKE_CASE : List[str] = image_processing(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) # Test not batched input (PIL images) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = prepare_semantic_single_inputs() SCREAMING_SNAKE_CASE : Union[str, Any] = image_processing(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( 1, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) # Test batched input (PIL images) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = prepare_semantic_batch_inputs() SCREAMING_SNAKE_CASE : Any = image_processing(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( 2, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) def _lowerCAmelCase ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Any = prepare_semantic_single_inputs() SCREAMING_SNAKE_CASE : int = image_processing(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 150 ) SCREAMING_SNAKE_CASE : Optional[Any] = True SCREAMING_SNAKE_CASE : Any = image_processing(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 )
265
0
"""simple docstring""" import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _lowerCamelCase : Optional[int] = abspath(join(dirname(dirname(dirname(__file__))), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCAmelCase_ ( lowercase_ : Union[str, Any] ): '''simple docstring''' from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def lowerCAmelCase_ ( lowercase_ : Dict ): '''simple docstring''' from transformers.testing_utils import pytest_terminal_summary_main __SCREAMING_SNAKE_CASE : List[Any] = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ )
708
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class snake_case ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE_ ( self :int ): __SCREAMING_SNAKE_CASE : int = XLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) __SCREAMING_SNAKE_CASE : Tuple = torch.tensor([[0, 5_8_1, 1_0_2_6_9, 8_3, 9_9_9_4_2, 1_3_6, 6_0_7_4_2, 2_3, 7_0, 8_0_5_8_3, 1_8_2_7_6, 2]] ) # The dog is cute and lives in the garden house __SCREAMING_SNAKE_CASE : Any = torch.Size((1, 1_2, 7_6_8) ) # batch_size, sequence_length, embedding_vector_dim __SCREAMING_SNAKE_CASE : Union[str, Any] = torch.tensor( [[-0.0_1_0_1, 0.1_2_1_8, -0.0_8_0_3, 0.0_8_0_1, 0.1_3_2_7, 0.0_7_7_6, -0.1_2_1_5, 0.2_3_8_3, 0.3_3_3_8, 0.3_1_0_6, 0.0_3_0_0, 0.0_2_5_2]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __SCREAMING_SNAKE_CASE : List[str] = model(_lowerCamelCase )['''last_hidden_state'''].detach() self.assertEqual(output.shape , _lowerCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowerCamelCase , atol=1e-3 ) ) @slow def SCREAMING_SNAKE_CASE_ ( self :Optional[Any] ): __SCREAMING_SNAKE_CASE : Optional[Any] = XLMRobertaModel.from_pretrained('''xlm-roberta-large''' ) __SCREAMING_SNAKE_CASE : Optional[Any] = torch.tensor([[0, 5_8_1, 1_0_2_6_9, 8_3, 9_9_9_4_2, 1_3_6, 6_0_7_4_2, 2_3, 7_0, 8_0_5_8_3, 1_8_2_7_6, 2]] ) # The dog is cute and lives in the garden house __SCREAMING_SNAKE_CASE : List[str] = torch.Size((1, 1_2, 1_0_2_4) ) # batch_size, sequence_length, embedding_vector_dim __SCREAMING_SNAKE_CASE : str = torch.tensor( [[-0.0_6_9_9, -0.0_3_1_8, 0.0_7_0_5, -0.1_2_4_1, 0.0_9_9_9, -0.0_5_2_0, 0.1_0_0_4, -0.1_8_3_8, -0.4_7_0_4, 0.1_4_3_7, 0.0_8_2_1, 0.0_1_2_6]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __SCREAMING_SNAKE_CASE : List[str] = model(_lowerCamelCase )['''last_hidden_state'''].detach() self.assertEqual(output.shape , _lowerCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowerCamelCase , atol=1e-3 ) )
401
0
'''simple docstring''' import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch lowerCAmelCase : Tuple = 'sshleifer/bart-tiny-random' lowerCAmelCase : Tuple = 'patrickvonplaten/t5-tiny-random' @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase): @cached_property def UpperCAmelCase_ ( self )-> Any: '''simple docstring''' return AutoConfig.from_pretrained(A_ ) def UpperCAmelCase_ ( self )-> Any: '''simple docstring''' UpperCamelCase , *UpperCamelCase = create_student_by_copying_alternating_layers(A_ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def UpperCAmelCase_ ( self )-> str: '''simple docstring''' UpperCamelCase , *UpperCamelCase = create_student_by_copying_alternating_layers(A_ , tempfile.mkdtemp() , e=1 , d=A_ ) def UpperCAmelCase_ ( self )-> Optional[int]: '''simple docstring''' UpperCamelCase , *UpperCamelCase = create_student_by_copying_alternating_layers(A_ , tempfile.mkdtemp() , e=1 , d=A_ ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def UpperCAmelCase_ ( self )-> Optional[int]: '''simple docstring''' UpperCamelCase , *UpperCamelCase = create_student_by_copying_alternating_layers(A_ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def UpperCAmelCase_ ( self )-> int: '''simple docstring''' with self.assertRaises(A_ ): create_student_by_copying_alternating_layers(A_ , tempfile.mkdtemp() , e=A_ , d=A_ )
3
'''simple docstring''' import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor __A : Tuple = logging.get_logger(__name__) class _UpperCamelCase ( _A ): '''simple docstring''' def __init__( self , *_a , **_a ): """simple docstring""" warnings.warn( 'The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use OwlViTImageProcessor instead.' , _a , ) super().__init__(*_a , **_a )
394
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ : Union[str, Any] = logging.get_logger(__name__) lowercase__ : Union[str, Any] = { '''kssteven/ibert-roberta-base''': '''https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json''', '''kssteven/ibert-roberta-large''': '''https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json''', '''kssteven/ibert-roberta-large-mnli''': ( '''https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json''' ), } class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' lowerCAmelCase_ = '''ibert''' def __init__( self : Optional[Any] , __lowercase : Optional[Any]=3_05_22 , __lowercase : List[str]=7_68 , __lowercase : str=12 , __lowercase : str=12 , __lowercase : int=30_72 , __lowercase : List[str]="gelu" , __lowercase : List[str]=0.1 , __lowercase : List[Any]=0.1 , __lowercase : List[str]=5_12 , __lowercase : List[Any]=2 , __lowercase : List[Any]=0.02 , __lowercase : int=1E-12 , __lowercase : List[Any]=1 , __lowercase : int=0 , __lowercase : List[Any]=2 , __lowercase : Tuple="absolute" , __lowercase : Any=False , __lowercase : Union[str, Any]="none" , **__lowercase : str , ): """simple docstring""" super().__init__(pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = position_embedding_type snake_case_ = quant_mode snake_case_ = force_dequant class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' @property def snake_case__ ( self : Dict ): """simple docstring""" if self.task == "multiple-choice": snake_case_ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case_ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
721
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class UpperCAmelCase : '''simple docstring''' def __init__( self : List[Any] , __lowercase : Optional[Any] , __lowercase : Optional[int]=13 , __lowercase : str=7 , __lowercase : str=True , __lowercase : Optional[int]=True , __lowercase : Optional[int]=False , __lowercase : str=True , __lowercase : Optional[int]=99 , __lowercase : List[str]=32 , __lowercase : Tuple=5 , __lowercase : int=4 , __lowercase : Union[str, Any]=37 , __lowercase : Union[str, Any]="gelu" , __lowercase : Dict=0.1 , __lowercase : int=0.1 , __lowercase : Optional[Any]=5_12 , __lowercase : Any=16 , __lowercase : int=2 , __lowercase : Dict=0.02 , __lowercase : List[str]=3 , __lowercase : int=4 , __lowercase : str=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def snake_case__ ( self : Dict ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self : Any ): """simple docstring""" return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) def snake_case__ ( self : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : str , __lowercase : Dict , __lowercase : Any , __lowercase : Dict , __lowercase : Union[str, Any] ): """simple docstring""" snake_case_ = LlamaModel(config=__lowercase ) model.to(__lowercase ) model.eval() snake_case_ = model(__lowercase , attention_mask=__lowercase ) snake_case_ = model(__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : Tuple , __lowercase : Tuple , __lowercase : Optional[Any] , __lowercase : Any , __lowercase : Optional[int] , __lowercase : int , __lowercase : str , __lowercase : str , __lowercase : List[Any] , __lowercase : Optional[Any] , ): """simple docstring""" snake_case_ = True snake_case_ = LlamaModel(__lowercase ) model.to(__lowercase ) model.eval() snake_case_ = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , ) snake_case_ = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , ) snake_case_ = model(__lowercase , attention_mask=__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : Dict , __lowercase : Union[str, Any] , __lowercase : Any , __lowercase : Tuple , __lowercase : str , __lowercase : Dict , __lowercase : int , __lowercase : Optional[int] , __lowercase : int , __lowercase : str , ): """simple docstring""" snake_case_ = LlamaForCausalLM(config=__lowercase ) model.to(__lowercase ) model.eval() snake_case_ = model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self : str , __lowercase : List[Any] , __lowercase : Optional[Any] , __lowercase : Union[str, Any] , __lowercase : List[str] , __lowercase : List[str] , __lowercase : List[str] , __lowercase : Optional[int] , __lowercase : str , __lowercase : Tuple , ): """simple docstring""" snake_case_ = True snake_case_ = True snake_case_ = LlamaForCausalLM(config=__lowercase ) model.to(__lowercase ) model.eval() # first forward pass snake_case_ = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , use_cache=__lowercase , ) snake_case_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case_ = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case_ = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case_ = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , output_hidden_states=__lowercase , )["hidden_states"][0] snake_case_ = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , past_key_values=__lowercase , output_hidden_states=__lowercase , )["hidden_states"][0] # select random slice snake_case_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case_ = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-3 ) ) def snake_case__ ( self : Dict ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () lowerCAmelCase_ = (LlamaForCausalLM,) if is_torch_available() else () lowerCAmelCase_ = ( { '''feature-extraction''': LlamaModel, '''text-classification''': LlamaForSequenceClassification, '''text-generation''': LlamaForCausalLM, '''zero-shot''': LlamaForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False def snake_case__ ( self : int ): """simple docstring""" snake_case_ = LlamaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__lowercase , hidden_size=37 ) def snake_case__ ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def snake_case__ ( self : Optional[Any] ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def snake_case__ ( self : Optional[int] ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__lowercase ) def snake_case__ ( self : Dict ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = input_dict["input_ids"] snake_case_ = input_ids.ne(1 ).to(__lowercase ) snake_case_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) snake_case_ = LlamaForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() snake_case_ = model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case__ ( self : int ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = "single_label_classification" snake_case_ = input_dict["input_ids"] snake_case_ = input_ids.ne(1 ).to(__lowercase ) snake_case_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) snake_case_ = LlamaForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() snake_case_ = model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case__ ( self : str ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = "multi_label_classification" snake_case_ = input_dict["input_ids"] snake_case_ = input_ids.ne(1 ).to(__lowercase ) snake_case_ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) snake_case_ = LlamaForSequenceClassification(__lowercase ) model.to(__lowercase ) model.eval() snake_case_ = model(__lowercase , attention_mask=__lowercase , labels=__lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip("LLaMA buffers include complex numbers, which breaks this test" ) def snake_case__ ( self : Optional[Any] ): """simple docstring""" pass @parameterized.expand([("linear",), ("dynamic",)] ) def snake_case__ ( self : Any , __lowercase : Tuple ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ids_tensor([1, 10] , config.vocab_size ) snake_case_ = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights snake_case_ = LlamaModel(__lowercase ) original_model.to(__lowercase ) original_model.eval() snake_case_ = original_model(__lowercase ).last_hidden_state snake_case_ = original_model(__lowercase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights snake_case_ = {"type": scaling_type, "factor": 10.0} snake_case_ = LlamaModel(__lowercase ) scaled_model.to(__lowercase ) scaled_model.eval() snake_case_ = scaled_model(__lowercase ).last_hidden_state snake_case_ = scaled_model(__lowercase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(__lowercase , __lowercase , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(__lowercase , __lowercase , atol=1E-5 ) ) @require_torch class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!" ) @slow def snake_case__ ( self : Any ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf" , device_map="auto" ) snake_case_ = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 snake_case_ = torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]] ) torch.testing.assert_close(out.mean(-1 ) , __lowercase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off snake_case_ = torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __lowercase , atol=1E-5 , rtol=1E-5 ) @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!" ) @slow def snake_case__ ( self : Union[str, Any] ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-13b-hf" , device_map="auto" ) snake_case_ = model(torch.tensor(__lowercase ) ) # Expected mean on dim = -1 snake_case_ = torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]] ) torch.testing.assert_close(out.mean(-1 ) , __lowercase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off snake_case_ = torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __lowercase , atol=1E-5 , rtol=1E-5 ) @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!" ) @slow def snake_case__ ( self : Any ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-13b-chat-hf" , device_map="auto" ) snake_case_ = model(torch.tensor(__lowercase ) ) # Expected mean on dim = -1 snake_case_ = torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]] ) torch.testing.assert_close(out.mean(-1 ) , __lowercase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off snake_case_ = torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , __lowercase , atol=1E-2 , rtol=1E-2 ) @unittest.skip( "Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test" ) @slow def snake_case__ ( self : Tuple ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-70b-hf" , device_map="auto" ) snake_case_ = model(torch.tensor(__lowercase ) ) snake_case_ = torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , __lowercase , atol=1E-2 , rtol=1E-2 ) # fmt: off snake_case_ = torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __lowercase , atol=1E-5 , rtol=1E-5 ) @unittest.skip("Model is curently gated" ) @slow def snake_case__ ( self : str ): """simple docstring""" snake_case_ = "Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the \"princi" snake_case_ = "Simply put, the theory of relativity states that " snake_case_ = LlamaTokenizer.from_pretrained("meta-llama/Llama-2-13b-chat-hf" ) snake_case_ = tokenizer.encode(__lowercase , return_tensors="pt" ) snake_case_ = LlamaForCausalLM.from_pretrained( "meta-llama/Llama-2-13b-chat-hf" , device_map="sequential" , use_safetensors=__lowercase ) # greedy generation outputs snake_case_ = model.generate(__lowercase , max_new_tokens=64 , top_p=__lowercase , temperature=1 , do_sample=__lowercase ) snake_case_ = tokenizer.decode(generated_ids[0] , skip_special_tokens=__lowercase ) self.assertEqual(__lowercase , __lowercase )
139
0
"""simple docstring""" import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() __lowerCamelCase = logging.get_logger(__name__) def a ( __UpperCAmelCase : Dict ) -> Dict: __magic_name__: List[Any] = torch.load(__UpperCAmelCase , map_location="""cpu""" ) if "model" in sd.keys(): __magic_name__: str = torch.load(__UpperCAmelCase , map_location="""cpu""" )["""model"""] # pop unnecessary weights __magic_name__: Any = [ """decoder.version""", """decoder.output_projection.weight""", ] for key in keys_to_delete: if key in sd: sd.pop(__UpperCAmelCase ) __magic_name__: List[Any] = { """decoder.project_in_dim.weight""": """decoder.project_in.weight""", """decoder.project_out_dim.weight""": """decoder.project_out.weight""", """decoder.layer_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.layer_norm.bias""": """decoder.final_layer_norm.bias""", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __magic_name__: Tuple = sd.pop(__UpperCAmelCase ) __magic_name__: str = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __magic_name__: List[str] = sd[key] # We split QKV in separate Q,K,V __magic_name__: Union[str, Any] = key.replace(""".qkv_proj.""" , """.q_proj.""" ) __magic_name__: int = key.replace(""".qkv_proj.""" , """.k_proj.""" ) __magic_name__: List[str] = key.replace(""".qkv_proj.""" , """.v_proj.""" ) __magic_name__: str = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __magic_name__, __magic_name__, __magic_name__: int = torch.split(__UpperCAmelCase , depth // 3 , dim=0 ) __magic_name__: int = q __magic_name__: Optional[Any] = k __magic_name__: str = v del sd[key] return sd @torch.no_grad() def a ( __UpperCAmelCase : List[str] , __UpperCAmelCase : int , __UpperCAmelCase : int=None ) -> Any: __magic_name__: str = load_checkpoint(__UpperCAmelCase ) if config is not None: __magic_name__: Tuple = OPTConfig.from_pretrained(__UpperCAmelCase ) else: __magic_name__: List[Any] = OPTConfig() __magic_name__: Tuple = OPTModel(__UpperCAmelCase ).half().eval() model.load_state_dict(__UpperCAmelCase ) # Check results Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) model.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--fairseq_path', type=str, help=( 'path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:' ' https://huggingface.co/models?other=opt_metasq' ), ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--hf_config', default=None, type=str, help='Define HF config.') __lowerCamelCase = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
96
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class lowerCamelCase: '''simple docstring''' def __init__( self , snake_case_ , ): _A = parent _A = 13 _A = 7 _A = True _A = True _A = True _A = 99 _A = 32 _A = 2 _A = 4 _A = 37 _A = 'gelu' _A = 0.1 _A = 0.1 _A = 512 _A = 16 _A = 2 _A = 0.02 _A = 3 _A = 4 _A = None def lowerCAmelCase__ ( self ): _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_input_mask: _A = random_attention_mask([self.batch_size, self.seq_length] ) _A = None _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A = ids_tensor([self.batch_size] , self.num_choices ) _A = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase__ ( self ): ( ( _A ), ( _A ), ( _A ), ( _A ), ( _A ), ( _A ), ) = self.prepare_config_and_inputs() _A = True _A = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _A = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def lowerCAmelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _A = TFEsmModel(config=snake_case_ ) _A = {'input_ids': input_ids, 'attention_mask': input_mask} _A = model(snake_case_ ) _A = [input_ids, input_mask] _A = model(snake_case_ ) _A = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): _A = True _A = TFEsmModel(config=snake_case_ ) _A = { 'input_ids': input_ids, 'attention_mask': input_mask, 'encoder_hidden_states': encoder_hidden_states, 'encoder_attention_mask': encoder_attention_mask, } _A = model(snake_case_ ) _A = [input_ids, input_mask] _A = model(snake_case_ , encoder_hidden_states=snake_case_ ) # Also check the case where encoder outputs are not passed _A = model(snake_case_ , attention_mask=snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _A = TFEsmForMaskedLM(config=snake_case_ ) _A = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _A = self.num_labels _A = TFEsmForTokenClassification(config=snake_case_ ) _A = {'input_ids': input_ids, 'attention_mask': input_mask} _A = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase__ ( self ): _A = self.prepare_config_and_inputs() ( ( _A ), ( _A ), ( _A ), ( _A ), ( _A ), ( _A ), ) = config_and_inputs _A = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class lowerCamelCase( __snake_case , __snake_case , unittest.TestCase ): '''simple docstring''' __magic_name__ = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) __magic_name__ = ( { 'feature-extraction': TFEsmModel, 'fill-mask': TFEsmForMaskedLM, 'text-classification': TFEsmForSequenceClassification, 'token-classification': TFEsmForTokenClassification, 'zero-shot': TFEsmForSequenceClassification, } if is_tf_available() else {} ) __magic_name__ = False __magic_name__ = False def lowerCAmelCase__ ( self ): _A = TFEsmModelTester(self ) _A = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCAmelCase__ ( self ): self.config_tester.run_common_tests() def lowerCAmelCase__ ( self ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCAmelCase__ ( self ): _A = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*snake_case_ ) def lowerCAmelCase__ ( self ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case_ ) def lowerCAmelCase__ ( self ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case_ ) @slow def lowerCAmelCase__ ( self ): for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = TFEsmModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @unittest.skip('Protein models do not support embedding resizing.' ) def lowerCAmelCase__ ( self ): pass @unittest.skip('Protein models do not support embedding resizing.' ) def lowerCAmelCase__ ( self ): pass def lowerCAmelCase__ ( self ): _A, _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(snake_case_ ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer _A = model.get_bias() assert isinstance(snake_case_ , snake_case_ ) for k, v in name.items(): assert isinstance(snake_case_ , tf.Variable ) else: _A = model.get_output_embeddings() assert x is None _A = model.get_bias() assert name is None @require_tf class lowerCamelCase( unittest.TestCase ): '''simple docstring''' @slow def lowerCAmelCase__ ( self ): _A = TFEsmForMaskedLM.from_pretrained('facebook/esm2_t6_8M_UR50D' ) _A = tf.constant([[0, 1, 2, 3, 4, 5]] ) _A = model(snake_case_ )[0] _A = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , snake_case_ ) # compare the actual values for a slice. _A = tf.constant( [ [ [8.92_1518, -10.58_9814, -6.467_1307], [-6.396_7156, -13.91_1377, -1.121_1915], [-7.78_1247, -13.95_1557, -3.74_0592], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-2 ) ) @slow def lowerCAmelCase__ ( self ): _A = TFEsmModel.from_pretrained('facebook/esm2_t6_8M_UR50D' ) _A = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) _A = model(snake_case_ )[0] # compare the actual values for a slice. _A = tf.constant( [ [ [0.1444_3092, 0.5412_5327, 0.324_7739], [0.3034_0484, 0.0052_6676, 0.3107_7722], [0.3227_8043, -0.2498_7096, 0.341_4628], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
27
0
"""simple docstring""" from abc import ABC, abstractmethod from typing import List, Optional class snake_case ( __lowercase ): def __init__(self ): """simple docstring""" self.test() def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = False while not completed: if counter == 1: self.reset() SCREAMING_SNAKE_CASE_ = self.advance() if not self.does_advance(SCREAMING_SNAKE_CASE_ ): raise Exception( '''Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.''' ) SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = self.update(SCREAMING_SNAKE_CASE_ ) counter += 1 if counter > 1_00_00: raise Exception('''update() does not fulfill the constraint.''' ) if self.remaining() != 0: raise Exception('''Custom Constraint is not defined correctly.''' ) @abstractmethod def _lowercase (self ): """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def _lowercase (self ): """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def _lowercase (self ): """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) @abstractmethod def _lowercase (self , SCREAMING_SNAKE_CASE_=False ): """simple docstring""" raise NotImplementedError( f'{self.__class__} is an abstract class. Only classes inheriting this class can be called.' ) class snake_case ( __lowercase ): def __init__(self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" super(SCREAMING_SNAKE_CASE_ , self ).__init__() if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) or len(SCREAMING_SNAKE_CASE_ ) == 0: raise ValueError(f'`token_ids` has to be a non-empty list, but is {token_ids}.' ) if any((not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) or token_id < 0) for token_id in token_ids ): raise ValueError(f'Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.' ) SCREAMING_SNAKE_CASE_ = token_ids SCREAMING_SNAKE_CASE_ = len(self.token_ids ) SCREAMING_SNAKE_CASE_ = -1 # the index of the currently fulfilled step SCREAMING_SNAKE_CASE_ = False def _lowercase (self ): """simple docstring""" if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError(f'`token_id` has to be an `int`, but is {token_id} of type {type(SCREAMING_SNAKE_CASE_ )}' ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError(f'`token_id` has to be an `int`, but is {token_id} of type {type(SCREAMING_SNAKE_CASE_ )}' ) SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = False if self.does_advance(SCREAMING_SNAKE_CASE_ ): self.fulfilled_idx += 1 SCREAMING_SNAKE_CASE_ = True if self.fulfilled_idx == (self.seqlen - 1): SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = completed else: # failed to make progress. SCREAMING_SNAKE_CASE_ = True self.reset() return stepped, completed, reset def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = 0 def _lowercase (self ): """simple docstring""" return self.seqlen - (self.fulfilled_idx + 1) def _lowercase (self , SCREAMING_SNAKE_CASE_=False ): """simple docstring""" SCREAMING_SNAKE_CASE_ = PhrasalConstraint(self.token_ids ) if stateful: SCREAMING_SNAKE_CASE_ = self.seqlen SCREAMING_SNAKE_CASE_ = self.fulfilled_idx SCREAMING_SNAKE_CASE_ = self.completed return new_constraint class snake_case : def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=True ): """simple docstring""" SCREAMING_SNAKE_CASE_ = max([len(SCREAMING_SNAKE_CASE_ ) for one in nested_token_ids] ) SCREAMING_SNAKE_CASE_ = {} for token_ids in nested_token_ids: SCREAMING_SNAKE_CASE_ = root for tidx, token_id in enumerate(SCREAMING_SNAKE_CASE_ ): if token_id not in level: SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = level[token_id] if no_subsets and self.has_subsets(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError( '''Each list in `nested_token_ids` can\'t be a complete subset of another list, but is''' f' {nested_token_ids}.' ) SCREAMING_SNAKE_CASE_ = root def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.trie for current_token in current_seq: SCREAMING_SNAKE_CASE_ = start[current_token] SCREAMING_SNAKE_CASE_ = list(start.keys() ) return next_tokens def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.next_tokens(SCREAMING_SNAKE_CASE_ ) return len(SCREAMING_SNAKE_CASE_ ) == 0 def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = list(root.values() ) if len(SCREAMING_SNAKE_CASE_ ) == 0: return 1 else: return sum([self.count_leaves(SCREAMING_SNAKE_CASE_ ) for nn in next_nodes] ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.count_leaves(SCREAMING_SNAKE_CASE_ ) return len(SCREAMING_SNAKE_CASE_ ) != leaf_count class snake_case ( __lowercase ): def __init__(self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" super(SCREAMING_SNAKE_CASE_ , self ).__init__() if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) or len(SCREAMING_SNAKE_CASE_ ) == 0: raise ValueError(f'`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.' ) if any(not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for token_ids in nested_token_ids ): raise ValueError(f'`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.' ) if any( any((not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( f'Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.' ) SCREAMING_SNAKE_CASE_ = DisjunctiveTrie(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = nested_token_ids SCREAMING_SNAKE_CASE_ = self.trie.max_height SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = False def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.trie.next_tokens(self.current_seq ) if len(SCREAMING_SNAKE_CASE_ ) == 0: return None else: return token_list def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError(f'`token_id` is supposed to be type `int`, but is {token_id} of type {type(SCREAMING_SNAKE_CASE_ )}' ) SCREAMING_SNAKE_CASE_ = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError(f'`token_id` is supposed to be type `int`, but is {token_id} of type {type(SCREAMING_SNAKE_CASE_ )}' ) SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = False if self.does_advance(SCREAMING_SNAKE_CASE_ ): self.current_seq.append(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = True else: SCREAMING_SNAKE_CASE_ = True self.reset() SCREAMING_SNAKE_CASE_ = self.trie.reached_leaf(self.current_seq ) SCREAMING_SNAKE_CASE_ = completed return stepped, completed, reset def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = [] def _lowercase (self ): """simple docstring""" if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def _lowercase (self , SCREAMING_SNAKE_CASE_=False ): """simple docstring""" SCREAMING_SNAKE_CASE_ = DisjunctiveConstraint(self.token_ids ) if stateful: SCREAMING_SNAKE_CASE_ = self.seqlen SCREAMING_SNAKE_CASE_ = self.current_seq SCREAMING_SNAKE_CASE_ = self.completed return new_constraint class snake_case : def __init__(self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = constraints # max # of steps required to fulfill a given constraint SCREAMING_SNAKE_CASE_ = max([c.seqlen for c in constraints] ) SCREAMING_SNAKE_CASE_ = len(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = False self.init_state() def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = [constraint.copy(stateful=SCREAMING_SNAKE_CASE_ ) for constraint in self.constraints] def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" SCREAMING_SNAKE_CASE_ = constraint.advance() if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): token_list.append(SCREAMING_SNAKE_CASE_ ) elif isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): token_list.extend(SCREAMING_SNAKE_CASE_ ) else: SCREAMING_SNAKE_CASE_ = self.inprogress_constraint.advance() if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): token_list.append(SCREAMING_SNAKE_CASE_ ) elif isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): token_list.extend(SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) == 0: return None else: return token_list def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = self.add(SCREAMING_SNAKE_CASE_ ) # the entire list of constraints are fulfilled if self.completed: break def _lowercase (self , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise ValueError(f'`token_id` should be an `int`, but is `{token_id}`.' ) SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = False, False if self.completed: SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = self.inprogress_constraint.update(SCREAMING_SNAKE_CASE_ ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=SCREAMING_SNAKE_CASE_ ) ) SCREAMING_SNAKE_CASE_ = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) SCREAMING_SNAKE_CASE_ = None if len(self.pending_constraints ) == 0: # we're done! SCREAMING_SNAKE_CASE_ = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = pending_constraint.update(SCREAMING_SNAKE_CASE_ ) if not stepped: raise Exception( '''`constraint.update(token_id)` is not yielding incremental progress, ''' '''even though `constraint.does_advance(token_id)` is true.''' ) if complete: self.complete_constraints.append(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = None if not complete and stepped: SCREAMING_SNAKE_CASE_ = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". SCREAMING_SNAKE_CASE_ = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. SCREAMING_SNAKE_CASE_ = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def _lowercase (self , SCREAMING_SNAKE_CASE_=True ): """simple docstring""" SCREAMING_SNAKE_CASE_ = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: SCREAMING_SNAKE_CASE_ = [ constraint.copy(stateful=SCREAMING_SNAKE_CASE_ ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: SCREAMING_SNAKE_CASE_ = self.inprogress_constraint.copy(stateful=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = [constraint.copy() for constraint in self.pending_constraints] return new_state
710
"""simple docstring""" from collections.abc import Sequence from queue import Queue class snake_case : def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None ): """simple docstring""" SCREAMING_SNAKE_CASE_ = start SCREAMING_SNAKE_CASE_ = end SCREAMING_SNAKE_CASE_ = val SCREAMING_SNAKE_CASE_ = (start + end) // 2 SCREAMING_SNAKE_CASE_ = left SCREAMING_SNAKE_CASE_ = right def __repr__(self ): """simple docstring""" return f'SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})' class snake_case : def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = collection SCREAMING_SNAKE_CASE_ = function if self.collection: SCREAMING_SNAKE_CASE_ = self._build_tree(0 , len(SCREAMING_SNAKE_CASE_ ) - 1 ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" self._update_tree(self.root , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" return self._query_range(self.root , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if start == end: return SegmentTreeNode(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , self.collection[start] ) SCREAMING_SNAKE_CASE_ = (start + end) // 2 SCREAMING_SNAKE_CASE_ = self._build_tree(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = self._build_tree(mid + 1 , SCREAMING_SNAKE_CASE_ ) return SegmentTreeNode(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , self.fn(left.val , right.val ) , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if node.start == i and node.end == i: SCREAMING_SNAKE_CASE_ = val return if i <= node.mid: self._update_tree(node.left , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else: self._update_tree(node.right , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE_ = self.fn(node.left.val , node.right.val ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , SCREAMING_SNAKE_CASE_ , node.mid ) , self._query_range(node.right , node.mid + 1 , SCREAMING_SNAKE_CASE_ ) , ) else: # range in right child tree return self._query_range(node.right , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def _lowercase (self ): """simple docstring""" if self.root is not None: SCREAMING_SNAKE_CASE_ = Queue() queue.put(self.root ) while not queue.empty(): SCREAMING_SNAKE_CASE_ = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('*' * 50) lowerCAmelCase__ = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
628
0
from scipy.stats import pearsonr import datasets SCREAMING_SNAKE_CASE__ : str = """ Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ SCREAMING_SNAKE_CASE__ : Tuple = """ Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ SCREAMING_SNAKE_CASE__ : int = """ @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): def A ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'''] , ) def A ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ): """simple docstring""" if return_pvalue: __magic_name__ :str = pearsonr(__lowerCAmelCase , __lowerCAmelCase ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__lowerCAmelCase , __lowerCAmelCase )[0] )}
0
import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin A__ = ''' Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] ''' class a ( unittest.TestCase , __lowerCamelCase ): def __lowerCamelCase ( self :str ): snake_case__ : Union[str, Any] = load_tool('''text-question-answering''' ) self.tool.setup() snake_case__ : int = load_tool('''text-question-answering''' ,remote=__lowercase ) def __lowerCamelCase ( self :List[Any] ): snake_case__ : int = self.tool(__lowercase ,'''What did Hugging Face do in April 2021?''' ) self.assertEqual(__lowercase ,'''launched the BigScience Research Workshop''' ) def __lowerCamelCase ( self :Optional[int] ): snake_case__ : int = self.remote_tool(__lowercase ,'''What did Hugging Face do in April 2021?''' ) self.assertEqual(__lowercase ,'''launched the BigScience Research Workshop''' ) def __lowerCamelCase ( self :Union[str, Any] ): snake_case__ : int = self.tool(text=__lowercase ,question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(__lowercase ,'''launched the BigScience Research Workshop''' ) def __lowerCamelCase ( self :Optional[Any] ): snake_case__ : Optional[Any] = self.remote_tool(text=__lowercase ,question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(__lowercase ,'''launched the BigScience Research Workshop''' )
252
0
import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging snake_case__ = logging.get_logger(__name__) class UpperCamelCase ( __lowercase ): '''simple docstring''' A_ = ['input_features'] def __init__( self , A_=80 , A_=1_60_00 , A_=1_60 , A_=30 , A_=4_00 , A_=0.0 , A_=False , **A_ , ) -> Union[str, Any]: """simple docstring""" super().__init__( feature_size=A_ , sampling_rate=A_ , padding_value=A_ , return_attention_mask=A_ , **A_ , ) _lowerCamelCase = n_fft _lowerCamelCase = hop_length _lowerCamelCase = chunk_length _lowerCamelCase = chunk_length * sampling_rate _lowerCamelCase = self.n_samples // hop_length _lowerCamelCase = sampling_rate _lowerCamelCase = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=A_ , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=A_ , norm='''slaney''' , mel_scale='''slaney''' , ) def UpperCamelCase_ ( self , A_ ) -> np.ndarray: """simple docstring""" _lowerCamelCase = spectrogram( A_ , window_function(self.n_fft , '''hann''' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='''log10''' , ) _lowerCamelCase = log_spec[:, :-1] _lowerCamelCase = np.maximum(A_ , log_spec.max() - 8.0 ) _lowerCamelCase = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def UpperCamelCase_ ( A_ , A_ , A_ = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: _lowerCamelCase = np.array(A_ , np.intaa ) _lowerCamelCase = [] for vector, length in zip(A_ , attention_mask.sum(-1 ) ): _lowerCamelCase = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _lowerCamelCase = padding_value normed_input_values.append(A_ ) else: _lowerCamelCase = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self , A_ , A_ = True , A_ = None , A_ = None , A_ = None , A_ = "max_length" , A_ = None , A_ = None , A_ = None , **A_ , ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' F' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' F' was sampled with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) _lowerCamelCase = isinstance(A_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}' ) _lowerCamelCase = is_batched_numpy or ( isinstance(A_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _lowerCamelCase = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(A_ , np.ndarray ): _lowerCamelCase = np.asarray(A_ , dtype=np.floataa ) elif isinstance(A_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _lowerCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _lowerCamelCase = [np.asarray([raw_speech] ).T] _lowerCamelCase = BatchFeature({'''input_features''': raw_speech} ) # convert into correct format for padding _lowerCamelCase = self.pad( A_ , padding=A_ , max_length=max_length if max_length else self.n_samples , truncation=A_ , pad_to_multiple_of=A_ , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: _lowerCamelCase = self.zero_mean_unit_var_norm( padded_inputs['''input_features'''] , attention_mask=padded_inputs['''attention_mask'''] , padding_value=self.padding_value , ) _lowerCamelCase = np.stack(padded_inputs['''input_features'''] , axis=0 ) # make sure list is in array format _lowerCamelCase = padded_inputs.get('''input_features''' ).transpose(2 , 0 , 1 ) _lowerCamelCase = [self._np_extract_fbank_features(A_ ) for waveform in input_features[0]] if isinstance(input_features[0] , A_ ): _lowerCamelCase = [np.asarray(A_ , dtype=np.floataa ) for feature in input_features] else: _lowerCamelCase = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) _lowerCamelCase = padded_inputs['''attention_mask'''][:, :: self.hop_length] if return_tensors is not None: _lowerCamelCase = padded_inputs.convert_to_tensors(A_ ) return padded_inputs def UpperCamelCase_ ( self ) -> Dict[str, Any]: """simple docstring""" _lowerCamelCase = copy.deepcopy(self.__dict__ ) _lowerCamelCase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
701
def __magic_name__( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __magic_name__( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' _lowerCamelCase = 0 while b > 0: if b & 1: _lowerCamelCase = ((res % c) + (a % c)) % c a += a b >>= 1 return res
638
0
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer SCREAMING_SNAKE_CASE__ : Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Any = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__ : Optional[int] = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } SCREAMING_SNAKE_CASE__ : Any = { """allenai/led-base-16384""": 1_63_84, } class lowerCamelCase_ ( lowerCAmelCase__ ): a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = LEDTokenizer a__ = ['''input_ids''', '''attention_mask'''] def __init__( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase="replace" , __lowerCAmelCase="<s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="<s>" , __lowerCAmelCase="<unk>" , __lowerCAmelCase="<pad>" , __lowerCAmelCase="<mask>" , __lowerCAmelCase=False , __lowerCAmelCase=True , **__lowerCAmelCase , ): """simple docstring""" super().__init__( __UpperCAmelCase , __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , errors=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , add_prefix_space=__UpperCAmelCase , trim_offsets=__UpperCAmelCase , **__UpperCAmelCase , ) __magic_name__ :Any = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , __UpperCAmelCase ) != add_prefix_space: __magic_name__ :Union[str, Any] = getattr(__UpperCAmelCase , pre_tok_state.pop('''type''' ) ) __magic_name__ :int = add_prefix_space __magic_name__ :int = pre_tok_class(**__UpperCAmelCase ) __magic_name__ :str = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` __magic_name__ :Tuple = '''post_processor''' __magic_name__ :Tuple = getattr(self.backend_tokenizer , __UpperCAmelCase , __UpperCAmelCase ) if tokenizer_component_instance: __magic_name__ :str = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __magic_name__ :Any = tuple(state['''sep'''] ) if "cls" in state: __magic_name__ :Optional[Any] = tuple(state['''cls'''] ) __magic_name__ :List[Any] = False if state.get('''add_prefix_space''' , __UpperCAmelCase ) != add_prefix_space: __magic_name__ :Optional[Any] = add_prefix_space __magic_name__ :Optional[int] = True if state.get('''trim_offsets''' , __UpperCAmelCase ) != trim_offsets: __magic_name__ :List[str] = trim_offsets __magic_name__ :List[Any] = True if changes_to_apply: __magic_name__ :Any = getattr(__UpperCAmelCase , state.pop('''type''' ) ) __magic_name__ :Optional[Any] = component_class(**__UpperCAmelCase ) setattr(self.backend_tokenizer , __UpperCAmelCase , __UpperCAmelCase ) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def A ( self ): """simple docstring""" if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def A ( self , __lowerCAmelCase ): """simple docstring""" __magic_name__ :List[str] = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else value __magic_name__ :Union[str, Any] = value def A ( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" __magic_name__ :Union[str, Any] = kwargs.get('''is_split_into_words''' , __UpperCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._batch_encode_plus(*__UpperCAmelCase , **__UpperCAmelCase ) def A ( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" __magic_name__ :str = kwargs.get('''is_split_into_words''' , __UpperCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' '''to use it with pretokenized inputs.''' ) return super()._encode_plus(*__UpperCAmelCase , **__UpperCAmelCase ) def A ( self , __lowerCAmelCase , __lowerCAmelCase = None ): """simple docstring""" __magic_name__ :Union[str, Any] = self._tokenizer.model.save(__UpperCAmelCase , name=__UpperCAmelCase ) return tuple(__UpperCAmelCase ) def A ( self , __lowerCAmelCase , __lowerCAmelCase=None ): """simple docstring""" __magic_name__ :List[str] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A ( self , __lowerCAmelCase , __lowerCAmelCase = None ): """simple docstring""" __magic_name__ :Tuple = [self.sep_token_id] __magic_name__ :Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A ( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = PaddingStrategy.DO_NOT_PAD , __lowerCAmelCase = None , __lowerCAmelCase = None , ): """simple docstring""" __magic_name__ :Dict = super()._pad( encoded_inputs=__UpperCAmelCase , max_length=__UpperCAmelCase , padding_strategy=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , ) # Load from model defaults if return_attention_mask is None: __magic_name__ :Tuple = '''attention_mask''' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: __magic_name__ :List[str] = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. __magic_name__ :Dict = len(encoded_inputs['''global_attention_mask'''] ) != len(__UpperCAmelCase ) if needs_to_be_padded: __magic_name__ :List[str] = len(__UpperCAmelCase ) - len(encoded_inputs['''global_attention_mask'''] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` __magic_name__ :Tuple = ( encoded_inputs['''global_attention_mask'''] + [-1] * difference ) elif self.padding_side == "left": __magic_name__ :str = [-1] * difference + encoded_inputs[ '''global_attention_mask''' ] else: raise ValueError('''Invalid padding strategy:''' + str(self.padding_side ) ) return encoded_inputs
0
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation a_ = logging.get_logger(__name__) a_ = {"""vocab_file""": """spiece.model"""} a_ = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } a_ = { """AI-Sweden/gpt-sw3-126m""": 2_048, """AI-Sweden/gpt-sw3-350m""": 2_048, """AI-Sweden/gpt-sw3-1.6b""": 2_048, """AI-Sweden/gpt-sw3-6.7b""": 2_048, """AI-Sweden/gpt-sw3-20b""": 2_048, } class __lowerCAmelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self , __UpperCAmelCase , __UpperCAmelCase=False , __UpperCAmelCase=False , __UpperCAmelCase=False , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs __lowerCamelCase = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) __lowerCamelCase = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __lowerCamelCase = '''<|endoftext|>''' if eos_token is None else eos_token __lowerCamelCase = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __lowerCamelCase = unk_token if pad_token is None else pad_token __lowerCamelCase = eos_token if bos_token is None else bos_token else: __lowerCamelCase = '''<pad>''' if pad_token is None else pad_token __lowerCamelCase = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) __lowerCamelCase = do_lower_case __lowerCamelCase = remove_space __lowerCamelCase = keep_accents __lowerCamelCase = vocab_file __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) # Used for whitespace normalization in input texts # fmt : off __lowerCamelCase = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __lowerCamelCase = re.compile( F"""[{"".join(map(__UpperCAmelCase , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]""" ) def __getstate__( self ): '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.non_printing_characters_re.sub('''''' , __UpperCAmelCase ) # Normalize whitespaces __lowerCamelCase = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization __lowerCamelCase = unicodedata.normalize('''NFC''' , __UpperCAmelCase ) return text def lowerCamelCase ( self , __UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.preprocess_text(__UpperCAmelCase ) return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.PieceToId(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.IdToPiece(__UpperCAmelCase ) @staticmethod def lowerCamelCase ( __UpperCAmelCase ): '''simple docstring''' return out_string def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = '''''' __lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCAmelCase ) + token __lowerCamelCase = True __lowerCamelCase = [] else: current_sub_tokens.append(__UpperCAmelCase ) __lowerCamelCase = False out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , '''wb''' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase = False ): '''simple docstring''' if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __lowerCamelCase = self.preprocess_text(__UpperCAmelCase ) __lowerCamelCase = self.sp_model.encode(__UpperCAmelCase ) else: __lowerCamelCase = [self.preprocess_text(__UpperCAmelCase ) for t in text] __lowerCamelCase = self.sp_model.encode(__UpperCAmelCase ) if return_tensors is True or return_tensors == "pt": __lowerCamelCase = torch.tensor(__UpperCAmelCase ) return token_ids def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' return self.sp_model.decode(__UpperCAmelCase ) def lowerCamelCase ( self , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = [F"""User: {text}""" if is_user else F"""Bot: {text}""" for is_user, text in conversation.iter_texts()] __lowerCamelCase = ( F"""{self.eos_token}{self.bos_token}""" + F"""{self.bos_token}""".join(__UpperCAmelCase ) + F"""{self.bos_token}Bot:""" ) return self.encode(text=__UpperCAmelCase )
175
0
from typing import List from .keymap import KEYMAP, get_character def UpperCAmelCase_ ( __UpperCamelCase ): def decorator(__UpperCamelCase ): SCREAMING_SNAKE_CASE__ =getattr(__UpperCamelCase, """handle_key""", [] ) handle += [key] setattr(__UpperCamelCase, """handle_key""", __UpperCamelCase ) return func return decorator def UpperCAmelCase_ ( *__UpperCamelCase ): def decorator(__UpperCamelCase ): SCREAMING_SNAKE_CASE__ =getattr(__UpperCamelCase, """handle_key""", [] ) handle += keys setattr(__UpperCamelCase, """handle_key""", __UpperCamelCase ) return func return decorator class __a ( __lowerCamelCase ): """simple docstring""" def __new__( cls : Any ,_UpperCamelCase : Optional[int] ,_UpperCamelCase : List[Any] ,_UpperCamelCase : Optional[int] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ =super().__new__(cls ,_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) if not hasattr(_UpperCamelCase ,"""key_handler""" ): setattr(_UpperCamelCase ,"""key_handler""" ,{} ) setattr(_UpperCamelCase ,"""handle_input""" ,KeyHandler.handle_input ) for value in attrs.values(): SCREAMING_SNAKE_CASE__ =getattr(_UpperCamelCase ,"""handle_key""" ,[] ) for key in handled_keys: SCREAMING_SNAKE_CASE__ =value return new_cls @staticmethod def __A ( cls : List[str] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE__ =get_character() if char != KEYMAP["undefined"]: SCREAMING_SNAKE_CASE__ =ord(_UpperCamelCase ) SCREAMING_SNAKE_CASE__ =cls.key_handler.get(_UpperCamelCase ) if handler: SCREAMING_SNAKE_CASE__ =char return handler(cls ) else: return None def UpperCAmelCase_ ( cls ): return KeyHandler(cls.__name__, cls.__bases__, cls.__dict__.copy() )
588
import warnings from functools import wraps from typing import Callable def UpperCAmelCase_ ( __UpperCamelCase ): @wraps(__UpperCamelCase ) def _inner_fn(*__UpperCamelCase, **__UpperCamelCase ): warnings.warn( (f"""'{fn.__name__}' is experimental and might be subject to breaking changes in the future."""), __UpperCamelCase, ) return fn(*__UpperCamelCase, **__UpperCamelCase ) return _inner_fn
588
1
'''simple docstring''' def _a ( _lowerCamelCase , _lowerCamelCase = False ) -> bool: """simple docstring""" if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3_3170_4406_4679_8873_8596_1981 and not allow_probable: raise ValueError( """Warning: upper bound of deterministic test is exceeded. """ """Pass allow_probable=True to allow probabilistic test. """ """A return value of True indicates a probable prime.""" ) # array bounds provided by analysis __snake_case : List[Any] = [ 2047, 137_3653, 2532_6001, 32_1503_1751, 2_1523_0289_8747, 3_4747_4966_0383, 341_5500_7172_8321, 1, 382_5123_0565_4641_3051, 1, 1, 3186_6585_7834_0311_5116_7461, 3_3170_4406_4679_8873_8596_1981, ] __snake_case : str = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(_lowerCamelCase , 1 ): if n < _p: # then we have our last prime to check __snake_case : Tuple = primes[:idx] break __snake_case , __snake_case : Union[str, Any] = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: __snake_case : Tuple = False for r in range(_lowerCamelCase ): __snake_case : Tuple = pow(_lowerCamelCase , d * 2**r , _lowerCamelCase ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): __snake_case : List[str] = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def _a ( ) -> None: """simple docstring""" assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(83_8201 ) assert miller_rabin(83_8207 ) # 1_373_653 assert not miller_rabin(1731_6001 ) assert miller_rabin(1731_6017 ) # 25_326_001 assert not miller_rabin(30_7838_6641 ) assert miller_rabin(30_7838_6653 ) # 3_215_031_751 assert not miller_rabin(1_7130_4557_4801 ) assert miller_rabin(1_7130_4557_4819 ) # 2_152_302_898_747 assert not miller_rabin(2_7797_9972_8307 ) assert miller_rabin(2_7797_9972_8327 ) # 3_474_749_660_383 assert not miller_rabin(113_8500_2390_9441 ) assert miller_rabin(113_8500_2390_9527 ) # 341_550_071_728_321 assert not miller_rabin(127_5041_0188_4880_4351 ) assert miller_rabin(127_5041_0188_4880_4391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(796_6646_4458_5077_8779_1867 ) assert miller_rabin(796_6646_4458_5077_8779_1951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(5528_4067_7446_6478_9766_0333 ) assert miller_rabin(5528_4067_7446_6478_9766_0359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
26
from __future__ import annotations def __lowerCamelCase ( __lowerCAmelCase : str , __lowerCAmelCase : Tuple , __lowerCAmelCase : int , __lowerCAmelCase : Tuple ) -> Optional[Any]: # noqa: E741 while r - l > 1: __UpperCamelCase : int = (l + r) // 2 if v[m] >= key: __UpperCamelCase : Dict = m else: __UpperCamelCase : Optional[Any] = m # noqa: E741 return r def __lowerCamelCase ( __lowerCAmelCase : list[int] ) -> int: if len(__lowerCAmelCase ) == 0: return 0 __UpperCamelCase : Optional[Any] = [0] * len(__lowerCAmelCase ) __UpperCamelCase : Optional[int] = 1 __UpperCamelCase : Union[str, Any] = v[0] for i in range(1 , len(__lowerCAmelCase ) ): if v[i] < tail[0]: __UpperCamelCase : Dict = v[i] elif v[i] > tail[length - 1]: __UpperCamelCase : Optional[int] = v[i] length += 1 else: __UpperCamelCase : List[Any] = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
269
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = { "configuration_speech_to_text": ["SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Speech2TextConfig"], "processing_speech_to_text": ["Speech2TextProcessor"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ["Speech2TextTokenizer"] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ["Speech2TextFeatureExtractor"] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ "TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFSpeech2TextForConditionalGeneration", "TFSpeech2TextModel", "TFSpeech2TextPreTrainedModel", ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ "SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "Speech2TextForConditionalGeneration", "Speech2TextModel", "Speech2TextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class a_ ( _lowerCAmelCase ): def __init__( self : Optional[int] , lowercase : NestedDataStructureLike[PathLike] , lowercase : Optional[NamedSplit] = None , lowercase : Optional[Features] = None , lowercase : str = None , lowercase : bool = False , lowercase : bool = False , lowercase : Optional[int] = None , **lowercase : List[Any] , ): """simple docstring""" super().__init__( lowercase , split=lowercase , features=lowercase , cache_dir=lowercase , keep_in_memory=lowercase , streaming=lowercase , num_proc=lowercase , **lowercase , ) lowercase_ :str = path_or_paths if isinstance(lowercase , lowercase ) else {self.split: path_or_paths} lowercase_ :Optional[int] = Text( cache_dir=lowercase , data_files=lowercase , features=lowercase , **lowercase , ) def lowercase__ ( self : Optional[int] ): """simple docstring""" if self.streaming: lowercase_ :str = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: lowercase_ :List[Any] = None lowercase_ :Any = None lowercase_ :Optional[int] = None lowercase_ :Optional[int] = None self.builder.download_and_prepare( download_config=lowercase , download_mode=lowercase , verification_mode=lowercase , base_path=lowercase , num_proc=self.num_proc , ) lowercase_ :int = self.builder.as_dataset( split=self.split , verification_mode=lowercase , in_memory=self.keep_in_memory ) return dataset
172
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def UpperCAmelCase_ ( __lowerCamelCase : int ): # vision encoder if "img_encoder.pos_embed" in name: lowercase_ :str = name.replace("img_encoder.pos_embed" ,"vision_model.embeddings.position_embeddings" ) if "img_encoder.patch_embed.proj" in name: lowercase_ :Optional[int] = name.replace("img_encoder.patch_embed.proj" ,"vision_model.embeddings.patch_embeddings.projection" ) if "img_encoder.patch_embed.norm" in name: lowercase_ :str = name.replace("img_encoder.patch_embed.norm" ,"vision_model.embeddings.layernorm" ) if "img_encoder.layers" in name: lowercase_ :Union[str, Any] = name.replace("img_encoder.layers" ,"vision_model.encoder.stages" ) if "blocks" in name and "res" not in name: lowercase_ :str = name.replace("blocks" ,"layers" ) if "attn" in name and "pre_assign" not in name: lowercase_ :Any = name.replace("attn" ,"self_attn" ) if "proj" in name and "self_attn" in name and "text" not in name: lowercase_ :List[Any] = name.replace("proj" ,"out_proj" ) if "pre_assign_attn.attn.proj" in name: lowercase_ :str = name.replace("pre_assign_attn.attn.proj" ,"pre_assign_attn.attn.out_proj" ) if "norm1" in name: lowercase_ :Optional[int] = name.replace("norm1" ,"layer_norm1" ) if "norm2" in name and "pre_assign" not in name: lowercase_ :Union[str, Any] = name.replace("norm2" ,"layer_norm2" ) if "img_encoder.norm" in name: lowercase_ :Tuple = name.replace("img_encoder.norm" ,"vision_model.layernorm" ) # text encoder if "text_encoder.token_embedding" in name: lowercase_ :Any = name.replace("text_encoder.token_embedding" ,"text_model.embeddings.token_embedding" ) if "text_encoder.positional_embedding" in name: lowercase_ :Optional[int] = name.replace("text_encoder.positional_embedding" ,"text_model.embeddings.position_embedding.weight" ) if "text_encoder.transformer.resblocks." in name: lowercase_ :Any = name.replace("text_encoder.transformer.resblocks." ,"text_model.encoder.layers." ) if "ln_1" in name: lowercase_ :int = name.replace("ln_1" ,"layer_norm1" ) if "ln_2" in name: lowercase_ :str = name.replace("ln_2" ,"layer_norm2" ) if "c_fc" in name: lowercase_ :str = name.replace("c_fc" ,"fc1" ) if "c_proj" in name: lowercase_ :int = name.replace("c_proj" ,"fc2" ) if "text_encoder" in name: lowercase_ :Union[str, Any] = name.replace("text_encoder" ,"text_model" ) if "ln_final" in name: lowercase_ :int = name.replace("ln_final" ,"final_layer_norm" ) # projection layers if "img_projector.linear_hidden." in name: lowercase_ :Optional[int] = name.replace("img_projector.linear_hidden." ,"visual_projection." ) if "img_projector.linear_out." in name: lowercase_ :Any = name.replace("img_projector.linear_out." ,"visual_projection.3." ) if "text_projector.linear_hidden" in name: lowercase_ :Optional[int] = name.replace("text_projector.linear_hidden" ,"text_projection" ) if "text_projector.linear_out" in name: lowercase_ :Any = name.replace("text_projector.linear_out" ,"text_projection.3" ) return name def UpperCAmelCase_ ( __lowerCamelCase : List[str] ,__lowerCamelCase : List[str] ): for key in orig_state_dict.copy().keys(): lowercase_ :List[Any] = orig_state_dict.pop(__lowerCamelCase ) if "qkv" in key: # weights and biases of the key, value and query projections of vision encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors lowercase_ :List[Any] = key.split("." ) lowercase_ , lowercase_ :Union[str, Any] = int(key_split[2] ), int(key_split[4] ) lowercase_ :Optional[Any] = config.vision_config.hidden_size if "weight" in key: lowercase_ :Any = val[:dim, :] lowercase_ :Union[str, Any] = val[dim : dim * 2, :] lowercase_ :Union[str, Any] = val[-dim:, :] else: lowercase_ :str = val[:dim] lowercase_ :Optional[int] = val[dim : dim * 2] lowercase_ :Union[str, Any] = val[-dim:] elif "in_proj" in key: # weights and biases of the key, value and query projections of text encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors lowercase_ :Optional[Any] = key.split("." ) lowercase_ :Union[str, Any] = int(key_split[3] ) lowercase_ :List[Any] = config.text_config.hidden_size if "weight" in key: lowercase_ :int = val[:dim, :] lowercase_ :str = val[ dim : dim * 2, : ] lowercase_ :str = val[-dim:, :] else: lowercase_ :List[str] = val[:dim] lowercase_ :Union[str, Any] = val[dim : dim * 2] lowercase_ :Dict = val[-dim:] else: lowercase_ :Optional[Any] = rename_key(__lowerCamelCase ) # squeeze if necessary if ( "text_projection.0" in new_name or "text_projection.3" in new_name or "visual_projection.0" in new_name or "visual_projection.3" in new_name ): lowercase_ :Dict = val.squeeze_() else: lowercase_ :Any = val return orig_state_dict def UpperCAmelCase_ ( ): lowercase_ :List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase_ :List[Any] = Image.open(requests.get(__lowerCamelCase ,stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase_ ( __lowerCamelCase : int ,__lowerCamelCase : str ,__lowerCamelCase : List[str]="groupvit-gcc-yfcc" ,__lowerCamelCase : List[str]=False ): lowercase_ :List[str] = GroupViTConfig() lowercase_ :Dict = GroupViTModel(__lowerCamelCase ).eval() lowercase_ :int = torch.load(__lowerCamelCase ,map_location="cpu" )["model"] lowercase_ :int = convert_state_dict(__lowerCamelCase ,__lowerCamelCase ) lowercase_ , lowercase_ :str = model.load_state_dict(__lowerCamelCase ,strict=__lowerCamelCase ) assert missing_keys == ["text_model.embeddings.position_ids"] assert (unexpected_keys == ["multi_label_logit_scale"]) or (len(__lowerCamelCase ) == 0) # verify result lowercase_ :List[Any] = CLIPProcessor.from_pretrained("openai/clip-vit-base-patch32" ) lowercase_ :str = prepare_img() lowercase_ :str = processor(text=["a photo of a cat", "a photo of a dog"] ,images=__lowerCamelCase ,padding=__lowerCamelCase ,return_tensors="pt" ) with torch.no_grad(): lowercase_ :Tuple = model(**__lowerCamelCase ) if model_name == "groupvit-gcc-yfcc": lowercase_ :Tuple = torch.tensor([[13.3_523, 6.3_629]] ) elif model_name == "groupvit-gcc-redcaps": lowercase_ :Any = torch.tensor([[16.1_873, 8.6_230]] ) else: raise ValueError(F'Model name {model_name} not supported.' ) assert torch.allclose(outputs.logits_per_image ,__lowerCamelCase ,atol=1e-3 ) processor.save_pretrained(__lowerCamelCase ) model.save_pretrained(__lowerCamelCase ) print("Successfully saved processor and model to" ,__lowerCamelCase ) if push_to_hub: print("Pushing to the hub..." ) processor.push_to_hub(__lowerCamelCase ,organization="nielsr" ) model.push_to_hub(__lowerCamelCase ,organization="nielsr" ) if __name__ == "__main__": lowerCAmelCase : List[str] =argparse.ArgumentParser() parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to dump the processor and PyTorch model.''' ) parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to GroupViT checkpoint''') parser.add_argument( '''--model_name''', default='''groupvit-gccy-fcc''', type=str, help='''Name of the model. Expecting either \'groupvit-gcc-yfcc\' or \'groupvit-gcc-redcaps\'''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model and processor to the 🤗 hub using the provided `model_name`.''', ) lowerCAmelCase : Union[str, Any] =parser.parse_args() convert_groupvit_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
172
1
"""simple docstring""" from math import factorial _lowerCAmelCase = {str(digit): factorial(digit) for digit in range(10)} def UpperCamelCase ( _A ) -> int: if not isinstance(_A , _A ): raise TypeError("""Parameter number must be int""" ) if number < 0: raise ValueError("""Parameter number must be greater than or equal to 0""" ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(_A ) ) def UpperCamelCase ( _A = 60 , _A = 1_000_000 ) -> int: if not isinstance(_A , _A ) or not isinstance(_A , _A ): raise TypeError("""Parameters chain_length and number_limit must be int""" ) if chain_length <= 0 or number_limit <= 0: raise ValueError( """Parameters chain_length and number_limit must be greater than 0""" ) # the counter for the chains with the exact desired length lowercase : Optional[Any] = 0 # the cached sizes of the previous chains lowercase : dict[int, int] = {} for start_chain_element in range(1 , _A ): # The temporary set will contain the elements of the chain lowercase : Union[str, Any] = set() lowercase : List[Any] = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. lowercase : Tuple = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(_A ) chain_set_length += 1 lowercase : str = digit_factorial_sum(_A ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] lowercase : List[Any] = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F'{solution()}')
348
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class UpperCamelCase (unittest.TestCase ): _SCREAMING_SNAKE_CASE : Optional[int] = MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING _SCREAMING_SNAKE_CASE : int = TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def __snake_case ( self :Any , __magic_name__ :Optional[int] , __magic_name__ :Tuple , __magic_name__ :List[str] ) ->Optional[Any]: lowercase : List[Any] = TextaTextGenerationPipeline(model=__magic_name__ , tokenizer=__magic_name__ ) return generator, ["Something to write", "Something else"] def __snake_case ( self :Tuple , __magic_name__ :List[Any] , __magic_name__ :int ) ->Optional[Any]: lowercase : Optional[Any] = generator("""Something there""" ) self.assertEqual(__magic_name__ , [{"""generated_text""": ANY(__magic_name__ )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]["""generated_text"""].startswith("""Something there""" ) ) lowercase : int = generator(["""This is great !""", """Something else"""] , num_return_sequences=2 , do_sample=__magic_name__ ) self.assertEqual( __magic_name__ , [ [{"""generated_text""": ANY(__magic_name__ )}, {"""generated_text""": ANY(__magic_name__ )}], [{"""generated_text""": ANY(__magic_name__ )}, {"""generated_text""": ANY(__magic_name__ )}], ] , ) lowercase : Dict = generator( ["""This is great !""", """Something else"""] , num_return_sequences=2 , batch_size=2 , do_sample=__magic_name__ ) self.assertEqual( __magic_name__ , [ [{"""generated_text""": ANY(__magic_name__ )}, {"""generated_text""": ANY(__magic_name__ )}], [{"""generated_text""": ANY(__magic_name__ )}, {"""generated_text""": ANY(__magic_name__ )}], ] , ) with self.assertRaises(__magic_name__ ): generator(4 ) @require_torch def __snake_case ( self :int ) ->Any: lowercase : Union[str, Any] = pipeline("""text2text-generation""" , model="""patrickvonplaten/t5-tiny-random""" , framework="""pt""" ) # do_sample=False necessary for reproducibility lowercase : List[Any] = generator("""Something there""" , do_sample=__magic_name__ ) self.assertEqual(__magic_name__ , [{"""generated_text""": """"""}] ) lowercase : Dict = 3 lowercase : Optional[Any] = generator( """Something there""" , num_return_sequences=__magic_name__ , num_beams=__magic_name__ , ) lowercase : Tuple = [ {"""generated_text""": """Beide Beide Beide Beide Beide Beide Beide Beide Beide"""}, {"""generated_text""": """Beide Beide Beide Beide Beide Beide Beide Beide"""}, {"""generated_text""": """"""}, ] self.assertEqual(__magic_name__ , __magic_name__ ) lowercase : Dict = generator("""This is a test""" , do_sample=__magic_name__ , num_return_sequences=2 , return_tensors=__magic_name__ ) self.assertEqual( __magic_name__ , [ {"""generated_token_ids""": ANY(torch.Tensor )}, {"""generated_token_ids""": ANY(torch.Tensor )}, ] , ) lowercase : List[Any] = generator.model.config.eos_token_id lowercase : Dict = """<pad>""" lowercase : Optional[Any] = generator( ["""This is a test""", """This is a second test"""] , do_sample=__magic_name__ , num_return_sequences=2 , batch_size=2 , return_tensors=__magic_name__ , ) self.assertEqual( __magic_name__ , [ [ {"""generated_token_ids""": ANY(torch.Tensor )}, {"""generated_token_ids""": ANY(torch.Tensor )}, ], [ {"""generated_token_ids""": ANY(torch.Tensor )}, {"""generated_token_ids""": ANY(torch.Tensor )}, ], ] , ) @require_tf def __snake_case ( self :Optional[int] ) ->List[str]: lowercase : Dict = pipeline("""text2text-generation""" , model="""patrickvonplaten/t5-tiny-random""" , framework="""tf""" ) # do_sample=False necessary for reproducibility lowercase : List[Any] = generator("""Something there""" , do_sample=__magic_name__ ) self.assertEqual(__magic_name__ , [{"""generated_text""": """"""}] )
348
1
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def lowerCamelCase_ ( ) -> Optional[int]: """simple docstring""" snake_case_ : Dict = ArgumentParser( description=( '''PyTorch TPU distributed training launch ''' '''helper utility that will spawn up ''' '''multiple distributed processes''' ) ) # Optional arguments for the launch helper parser.add_argument('''--num_cores''' , type=_UpperCamelCase , default=1 , help='''Number of TPU cores to use (1 or 8).''' ) # positional parser.add_argument( '''training_script''' , type=_UpperCamelCase , help=( '''The full path to the single TPU training ''' '''program/script to be launched in parallel, ''' '''followed by all the arguments for the ''' '''training script''' ) , ) # rest from the training program parser.add_argument('''training_script_args''' , nargs=_UpperCamelCase ) return parser.parse_args() def lowerCamelCase_ ( ) -> Union[str, Any]: """simple docstring""" snake_case_ : Dict = parse_args() # Import training_script as a module. snake_case_ : str = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) snake_case_ : str = script_fpath.stem snake_case_ : Optional[int] = importlib.import_module(_UpperCamelCase ) # Patch sys.argv snake_case_ : Any = [args.training_script] + args.training_script_args + ['''--tpu_num_cores''', str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
60
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class __lowerCAmelCase ( unittest.TestCase ): lowerCamelCase_ : Tuple = inspect.getfile(accelerate.test_utils ) lowerCamelCase_ : Optional[Any] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) lowerCamelCase_ : Union[str, Any] = ['''accelerate''', '''launch'''] lowerCamelCase_ : Tuple = Path.home() / '''.cache/huggingface/accelerate''' lowerCamelCase_ : Tuple = '''default_config.yaml''' lowerCamelCase_ : str = config_folder / config_file lowerCamelCase_ : List[Any] = config_folder / '''_default_config.yaml''' lowerCamelCase_ : Dict = Path('''tests/test_configs''' ) @classmethod def lowerCamelCase (cls ) -> Dict: '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def lowerCamelCase (cls ) -> Any: '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def lowerCamelCase (self ) -> Tuple: '''simple docstring''' snake_case_ : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] , env=os.environ.copy() ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' for config in sorted(self.test_config_path.glob('''**/*.yaml''' ) ): with self.subTest(config_file=__magic_name__ ): execute_subprocess_async( self.base_cmd + ['''--config_file''', str(__magic_name__ ), self.test_file_path] , env=os.environ.copy() ) def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' execute_subprocess_async(['''accelerate''', '''test'''] , env=os.environ.copy() ) class __lowerCAmelCase ( unittest.TestCase ): lowerCamelCase_ : List[str] = '''test-tpu''' lowerCamelCase_ : Dict = '''us-central1-a''' lowerCamelCase_ : Any = '''ls''' lowerCamelCase_ : Dict = ['''accelerate''', '''tpu-config'''] lowerCamelCase_ : Tuple = '''cd /usr/share''' lowerCamelCase_ : List[Any] = '''tests/test_samples/test_command_file.sh''' lowerCamelCase_ : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : int = run_command( self.cmd + ['''--command''', self.command, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug'''] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : Optional[int] = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/0_12_0.yaml''', '''--command''', self.command, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug''', ] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[str] = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--debug'''] , return_stdout=__magic_name__ ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : List[Any] = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--command''', self.command, '''--debug'''] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : Any = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/latest.yaml''', '''--command''', self.command, '''--command''', '''echo "Hello World"''', '''--debug''', ] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo "Hello World" --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : str = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--command_file''', self.command_file, '''--debug'''] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : Tuple = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/0_12_0.yaml''', '''--command_file''', self.command_file, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug''', ] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> Optional[int]: '''simple docstring''' snake_case_ : Any = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--install_accelerate''', '''--debug'''] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo "hello world"; echo "this is a second command" --worker all''' , __magic_name__ , ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Optional[Any] = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/latest.yaml''', '''--install_accelerate''', '''--accelerate_version''', '''12.0.0''', '''--debug''', ] , return_stdout=__magic_name__ , ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo "hello world"; echo "this is a second command" --worker all''' , __magic_name__ , )
60
1
'''simple docstring''' import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient _SCREAMING_SNAKE_CASE = WebClient(token=os.environ['''CI_SLACK_BOT_TOKEN''']) def _lowerCAmelCase ( lowerCamelCase_ : Any ): __lowercase = test_results.split(''' ''' ) __lowercase = 0 __lowercase = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. __lowercase = expressions[-2] if '''=''' in expressions[-1] else expressions[-1] for i, expression in enumerate(lowerCamelCase_ ): if "failed" in expression: failed += int(expressions[i - 1] ) if "passed" in expression: success += int(expressions[i - 1] ) return failed, success, time_spent def _lowerCAmelCase ( lowerCamelCase_ : Union[str, Any] ): __lowercase = {} __lowercase = None __lowercase = False for line in failures_short_lines.split('''\n''' ): if re.search(r'''_ \[doctest\]''' , lowerCamelCase_ ): __lowercase = True __lowercase = line.split(''' ''' )[2] elif in_error and not line.split(''' ''' )[0].isdigit(): __lowercase = line __lowercase = False return failures class __lowercase : '''simple docstring''' def __init__(self ,_lowerCamelCase ,_lowerCamelCase ) -> Any: '''simple docstring''' __lowercase = title __lowercase = doc_test_results['''time_spent'''].split(''',''' )[0] __lowercase = doc_test_results['''success'''] __lowercase = doc_test_results['''failures'''] __lowercase = self.n_success + self.n_failures # Failures and success of the modeling tests __lowercase = doc_test_results @property def _UpperCAmelCase (self ) -> str: '''simple docstring''' __lowercase = [self._time_spent] __lowercase = 0 for time in time_spent: __lowercase = time.split(''':''' ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(_lowerCamelCase ) == 1: __lowercase = [0, 0, time_parts[0]] __lowercase , __lowercase , __lowercase = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 3600 + minutes * 60 + seconds __lowercase , __lowercase , __lowercase = total_secs // 3600, (total_secs % 3600) // 60, total_secs % 60 return f"{int(_lowerCamelCase )}h{int(_lowerCamelCase )}m{int(_lowerCamelCase )}s" @property def _UpperCAmelCase (self ) -> Dict: '''simple docstring''' return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def _UpperCAmelCase (self ) -> Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": f"🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.", "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } @property def _UpperCAmelCase (self ) -> Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": ( f"There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in" f" {self.time}." ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } @property def _UpperCAmelCase (self ) -> Dict: '''simple docstring''' __lowercase = 40 __lowercase = {k: v['''failed'''] for k, v in doc_test_results.items() if isinstance(_lowerCamelCase ,_lowerCamelCase )} __lowercase = '''''' for category, failures in category_failures.items(): if len(_lowerCamelCase ) == 0: continue if report != "": report += "\n\n" report += f"*{category} failures*:".ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(_lowerCamelCase ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": f"The following examples had failures:\n\n\n{report}\n", }, } @property def _UpperCAmelCase (self ) -> str: '''simple docstring''' __lowercase = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(_lowerCamelCase ) @staticmethod def _UpperCAmelCase () -> List[str]: '''simple docstring''' __lowercase = [ { '''type''': '''section''', '''text''': { '''type''': '''plain_text''', '''text''': '''There was an issue running the tests.''', }, '''accessory''': { '''type''': '''button''', '''text''': {'''type''': '''plain_text''', '''text''': '''Check Action results''', '''emoji''': True}, '''url''': f"https://github.com/huggingface/transformers/actions/runs/{os.environ['GITHUB_RUN_ID']}", }, } ] print('''Sending the following payload''' ) print(json.dumps({'''blocks''': json.loads(_lowerCamelCase )} ) ) client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] ,text='''There was an issue running the tests.''' ,blocks=_lowerCamelCase ,) def _UpperCAmelCase (self ) -> Tuple: '''simple docstring''' print('''Sending the following payload''' ) print(json.dumps({'''blocks''': json.loads(self.payload )} ) ) __lowercase = f"{self.n_failures} failures out of {self.n_tests} tests," if self.n_failures else '''All tests passed.''' __lowercase = client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] ,blocks=self.payload ,text=_lowerCamelCase ,) def _UpperCAmelCase (self ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) -> Union[str, Any]: '''simple docstring''' __lowercase = '''''' for key, value in failures.items(): __lowercase = value[:200] + ''' [Truncated]''' if len(_lowerCamelCase ) > 250 else value failures_text += f"*{key}*\n_{value}_\n\n" __lowercase = job_name __lowercase = {'''type''': '''section''', '''text''': {'''type''': '''mrkdwn''', '''text''': text}} if job_link is not None: __lowercase = { '''type''': '''button''', '''text''': {'''type''': '''plain_text''', '''text''': '''GitHub Action job''', '''emoji''': True}, '''url''': job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def _UpperCAmelCase (self ) -> Any: '''simple docstring''' if self.thread_ts is None: raise ValueError('''Can only post reply if a post has been made.''' ) __lowercase = self.doc_test_results.pop('''job_link''' ) self.doc_test_results.pop('''failures''' ) self.doc_test_results.pop('''success''' ) self.doc_test_results.pop('''time_spent''' ) __lowercase = sorted(self.doc_test_results.items() ,key=lambda _lowerCamelCase : t[0] ) for job, job_result in sorted_dict: if len(job_result['''failures'''] ): __lowercase = f"*Num failures* :{len(job_result['failed'] )} \n" __lowercase = job_result['''failures'''] __lowercase = self.get_reply_blocks(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,text=_lowerCamelCase ) print('''Sending the following reply''' ) print(json.dumps({'''blocks''': blocks} ) ) client.chat_postMessage( channel=os.environ['''CI_SLACK_CHANNEL_ID_DAILY'''] ,text=f"Results for {job}" ,blocks=_lowerCamelCase ,thread_ts=self.thread_ts['''ts'''] ,) time.sleep(1 ) def _lowerCAmelCase ( ): __lowercase = os.environ['''GITHUB_RUN_ID'''] __lowercase = f"https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100" __lowercase = requests.get(lowerCamelCase_ ).json() __lowercase = {} try: jobs.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) __lowercase = math.ceil((result['''total_count'''] - 1_0_0) / 1_0_0 ) for i in range(lowerCamelCase_ ): __lowercase = requests.get(url + f"&page={i + 2}" ).json() jobs.update({job['''name''']: job['''html_url'''] for job in result['''jobs''']} ) return jobs except Exception as e: print('''Unknown error, could not fetch links.''' , lowerCamelCase_ ) return {} def _lowerCAmelCase ( lowerCamelCase_ : str ): __lowercase = {} if os.path.exists(lowerCamelCase_ ): __lowercase = os.listdir(lowerCamelCase_ ) for file in files: try: with open(os.path.join(lowerCamelCase_ , lowerCamelCase_ ) , encoding='''utf-8''' ) as f: __lowercase = f.read() except UnicodeDecodeError as e: raise ValueError(f"Could not open {os.path.join(lowerCamelCase_ , lowerCamelCase_ )}." ) from e return _artifact def _lowerCAmelCase ( ): class __lowercase : '''simple docstring''' def __init__(self ,_lowerCamelCase ) -> Dict: '''simple docstring''' __lowercase = name __lowercase = [] def __str__(self ) -> List[str]: '''simple docstring''' return self.name def _UpperCAmelCase (self ,_lowerCamelCase ) -> Dict: '''simple docstring''' self.paths.append({'''name''': self.name, '''path''': path} ) __lowercase = {} __lowercase = filter(os.path.isdir , os.listdir() ) for directory in directories: __lowercase = directory if artifact_name not in _available_artifacts: __lowercase = Artifact(lowerCamelCase_ ) _available_artifacts[artifact_name].add_path(lowerCamelCase_ ) return _available_artifacts if __name__ == "__main__": _SCREAMING_SNAKE_CASE = get_job_links() _SCREAMING_SNAKE_CASE = retrieve_available_artifacts() _SCREAMING_SNAKE_CASE = collections.OrderedDict( [ ('''*.py''', '''API Examples'''), ('''*.md''', '''MD Examples'''), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' _SCREAMING_SNAKE_CASE = { v: { '''failed''': [], '''failures''': {}, } for v in docs.values() } # Link to the GitHub Action job _SCREAMING_SNAKE_CASE = github_actions_job_links.get('''run_doctests''') _SCREAMING_SNAKE_CASE = available_artifacts['''doc_tests_gpu_test_reports'''].paths[0] _SCREAMING_SNAKE_CASE = retrieve_artifact(artifact_path['''name''']) if "stats" in artifact: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = handle_test_results(artifact['''stats''']) _SCREAMING_SNAKE_CASE = failed _SCREAMING_SNAKE_CASE = success _SCREAMING_SNAKE_CASE = time_spent[1:-1] + ''', ''' _SCREAMING_SNAKE_CASE = extract_first_line_failure(artifact['''failures_short''']) for line in artifact["summary_short"].split('''\n'''): if re.search('''FAILED''', line): _SCREAMING_SNAKE_CASE = line.replace('''FAILED ''', '''''') _SCREAMING_SNAKE_CASE = line.split()[0].replace('''\n''', '''''') if "::" in line: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = line.split('''::''') else: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): _SCREAMING_SNAKE_CASE = docs[file_regex] doc_test_results[category]["failed"].append(test) _SCREAMING_SNAKE_CASE = all_failures[test] if test in all_failures else '''N/A''' _SCREAMING_SNAKE_CASE = failure break _SCREAMING_SNAKE_CASE = Message('''🤗 Results of the doc tests.''', doc_test_results) message.post() message.post_reply()
56
'''simple docstring''' # flake8: noqa # Lint as: python3 _SCREAMING_SNAKE_CASE = [ '''VerificationMode''', '''Version''', '''disable_progress_bar''', '''enable_progress_bar''', '''is_progress_bar_enabled''', '''experimental''', ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_progress_bar_enabled from .version import Version from .experimental import experimental
56
1
'''simple docstring''' import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) UpperCAmelCase = logging.getLogger() UpperCAmelCase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __snake_case( __snake_case ): '''simple docstring''' def __snake_case ( self , A_ ) -> List[str]: os.makedirs(A_ , exist_ok=A_ ) lowerCAmelCase = {"source": "What is love ?", "target": "life"} lowerCAmelCase = {"train": 12, "val": 2, "test": 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: lowerCAmelCase = "\n".join([contents[field]] * n_lines[split] ) with open(os.path.join(A_ , f'{split}.{field}' ) , """w""" ) as f: f.write(A_ ) def __snake_case ( self , A_ , A_ = "pytorch" ) -> Tuple: lowerCAmelCase = self.get_auto_remove_tmp_dir() lowerCAmelCase = os.path.join(A_ , """output""" ) lowerCAmelCase = os.path.join(A_ , """data""" ) self._create_dummy_data(data_dir=A_ ) lowerCAmelCase = f'\n --data_dir {data_dir} \\n --output_dir {output_dir} \\n --model_name_or_path facebook/rag-sequence-base \\n --model_type rag_sequence \\n --do_train \\n --do_predict \\n --n_val -1 \\n --val_check_interval 1.0 \\n --train_batch_size 2 \\n --eval_batch_size 1 \\n --max_source_length 25 \\n --max_target_length 25 \\n --val_max_target_length 25 \\n --test_max_target_length 25 \\n --label_smoothing 0.1 \\n --dropout 0.1 \\n --attention_dropout 0.1 \\n --weight_decay 0.001 \\n --adam_epsilon 1e-08 \\n --max_grad_norm 0.1 \\n --lr_scheduler polynomial \\n --learning_rate 3e-04 \\n --num_train_epochs 1 \\n --warmup_steps 4 \\n --gradient_accumulation_steps 1 \\n --distributed-port 8787 \\n --use_dummy_dataset 1 \\n --distributed_retriever {distributed_retriever} \\n '.split() if gpus > 0: testargs.append(f'--gpus={gpus}' ) if is_apex_available(): testargs.append("""--fp16""" ) else: testargs.append("""--gpus=0""" ) testargs.append("""--distributed_backend=ddp_cpu""" ) testargs.append("""--num_processes=2""" ) lowerCAmelCase = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(A_ , env=self.get_env() ) lowerCAmelCase = os.path.join(A_ , """metrics.json""" ) with open(A_ ) as f: lowerCAmelCase = json.load(A_ ) return result @require_torch_gpu def __snake_case ( self ) -> List[Any]: lowerCAmelCase = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_multi_gpu def __snake_case ( self ) -> int: lowerCAmelCase = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_gpu @require_ray def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = self._run_finetune(gpus=1 , distributed_retriever="""ray""" ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 ) @require_torch_multi_gpu @require_ray def __snake_case ( self ) -> Optional[Any]: lowerCAmelCase = self._run_finetune(gpus=1 , distributed_retriever="""ray""" ) self.assertGreaterEqual(result["""test"""][0]["""test_avg_em"""] , 0.2 )
433
def A_ ( _lowerCAmelCase = 1000 ) -> int: UpperCamelCase : Optional[int] = -1 UpperCamelCase : int = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c UpperCamelCase : Optional[Any] = (n * n - 2 * a * n) // (2 * n - 2 * a) UpperCamelCase : List[str] = n - a - b if c * c == (a * a + b * b): UpperCamelCase : Union[str, Any] = a * b * c if candidate >= product: UpperCamelCase : Tuple = candidate return product if __name__ == "__main__": print(f"""{solution() = }""")
629
0
from __future__ import annotations def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Dict: """simple docstring""" if len(snake_case__ ) <= 1 or n <= 1: return insert_next(snake_case__ ,n - 1 ) rec_insertion_sort(snake_case__ ,n - 1 ) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" if index >= len(snake_case__ ) or collection[index - 1] <= collection[index]: return # Swaps adjacent elements since they are not in ascending order _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = ( collection[index], collection[index - 1], ) insert_next(snake_case__ ,index + 1 ) if __name__ == "__main__": UpperCamelCase = input('''Enter integers separated by spaces: ''') UpperCamelCase = [int(num) for num in numbers.split()] rec_insertion_sort(number_list, len(number_list)) print(number_list)
569
UpperCamelCase = 256 # Modulus to hash a string UpperCamelCase = 1_000_003 def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> bool: """simple docstring""" _SCREAMING_SNAKE_CASE = len(snake_case__ ) _SCREAMING_SNAKE_CASE = len(snake_case__ ) if p_len > t_len: return False _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 1 # Calculating the hash of pattern and substring of text for i in range(snake_case__ ): _SCREAMING_SNAKE_CASE = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus _SCREAMING_SNAKE_CASE = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue _SCREAMING_SNAKE_CASE = (modulus_power * alphabet_size) % modulus for i in range(0 ,t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash _SCREAMING_SNAKE_CASE = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def __lowerCamelCase ( ) -> None: """simple docstring""" _SCREAMING_SNAKE_CASE = """abc1abc12""" _SCREAMING_SNAKE_CASE = """alskfjaldsabc1abc1abc12k23adsfabcabc""" _SCREAMING_SNAKE_CASE = """alskfjaldsk23adsfabcabc""" assert rabin_karp(snake_case__ ,snake_case__ ) and not rabin_karp(snake_case__ ,snake_case__ ) # Test 2) _SCREAMING_SNAKE_CASE = """ABABX""" _SCREAMING_SNAKE_CASE = """ABABZABABYABABX""" assert rabin_karp(snake_case__ ,snake_case__ ) # Test 3) _SCREAMING_SNAKE_CASE = """AAAB""" _SCREAMING_SNAKE_CASE = """ABAAAAAB""" assert rabin_karp(snake_case__ ,snake_case__ ) # Test 4) _SCREAMING_SNAKE_CASE = """abcdabcy""" _SCREAMING_SNAKE_CASE = """abcxabcdabxabcdabcdabcy""" assert rabin_karp(snake_case__ ,snake_case__ ) # Test 5) _SCREAMING_SNAKE_CASE = """Lü""" _SCREAMING_SNAKE_CASE = """Lüsai""" assert rabin_karp(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = """Lue""" assert not rabin_karp(snake_case__ ,snake_case__ ) print("""Success.""" ) if __name__ == "__main__": test_rabin_karp()
569
1
"""simple docstring""" from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance A__ : Dict = 6378137.0 A__ : Dict = 6356752.314245 A__ : List[str] = 6_3_7_8_1_3_7 def _lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase: Any = (AXIS_A - AXIS_B) / AXIS_A # Parametric latitudes # https://en.wikipedia.org/wiki/Latitude#Parametric_(or_reduced)_latitude _lowercase: str = atan((1 - flattening) * tan(radians(_UpperCamelCase ) ) ) _lowercase: Dict = atan((1 - flattening) * tan(radians(_UpperCamelCase ) ) ) # Compute central angle between two points # using haversine theta. sigma = haversine_distance / equatorial radius _lowercase: Union[str, Any] = haversine_distance(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) / EQUATORIAL_RADIUS # Intermediate P and Q values _lowercase: List[Any] = (b_lata + b_lata) / 2 _lowercase: Any = (b_lata - b_lata) / 2 # Intermediate X value # X = (sigma - sin(sigma)) * sin^2Pcos^2Q / cos^2(sigma/2) _lowercase: Tuple = (sin(_UpperCamelCase ) ** 2) * (cos(_UpperCamelCase ) ** 2) _lowercase: Optional[int] = cos(sigma / 2 ) ** 2 _lowercase: List[Any] = (sigma - sin(_UpperCamelCase )) * (x_numerator / x_demonimator) # Intermediate Y value # Y = (sigma + sin(sigma)) * cos^2Psin^2Q / sin^2(sigma/2) _lowercase: List[Any] = (cos(_UpperCamelCase ) ** 2) * (sin(_UpperCamelCase ) ** 2) _lowercase: Dict = sin(sigma / 2 ) ** 2 _lowercase: int = (sigma + sin(_UpperCamelCase )) * (y_numerator / y_denominator) return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (x_value + y_value))) if __name__ == "__main__": import doctest doctest.testmod()
353
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def _lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase: int = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] _lowercase: Dict = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } _lowercase: Optional[Any] = f'''{src_lang}-{tgt_lang}''' _lowercase: Tuple = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(_UpperCamelCase , exist_ok=_UpperCamelCase ) _lowercase: List[str] = os.path.join(_UpperCamelCase , '''README.md''' ) print(f'''Generating {path}''' ) with open(_UpperCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(_UpperCamelCase ) # make sure we are under the root of the project A__ : List[str] = Path(__file__).resolve().parent.parent.parent A__ : Optional[int] = repo_dir / 'model_cards' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: A__ , A__ , A__ : int = model_name.split('-') A__ : Optional[int] = model_cards_dir / 'facebook' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
353
1
"""simple docstring""" from math import sqrt def a_ ( _lowerCAmelCase : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' must been an int and positive" lowercase__ : List[str] = True # 0 and 1 are none primes. if number <= 1: lowercase__ : int = False for divisor in range(2 , int(round(sqrt(_lowerCAmelCase ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowercase__ : List[str] = False break # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'status' must been from type bool" return status def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowercase__ : int = list(range(2 , n + 1 ) ) lowercase__ : Dict = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(_lowerCAmelCase ) ): for j in range(i + 1 , len(_lowerCAmelCase ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowercase__ : Tuple = 0 # filters actual prime numbers. lowercase__ : List[Any] = [x for x in begin_list if x != 0] # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n > 2), "'N' must been an int and > 2" lowercase__ : List[Any] = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(_lowerCAmelCase ): ans.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def a_ ( _lowerCAmelCase : str ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and number >= 0, "'number' must been an int and >= 0" lowercase__ : Any = [] # this list will be returns of the function. # potential prime number factors. lowercase__ : List[Any] = 2 lowercase__ : List[str] = number if number == 0 or number == 1: ans.append(_lowerCAmelCase ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(_lowerCAmelCase ): while quotient != 1: if is_prime(_lowerCAmelCase ) and (quotient % factor == 0): ans.append(_lowerCAmelCase ) quotient /= factor else: factor += 1 else: ans.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase__ : List[str] = 0 # prime factorization of 'number' lowercase__ : str = prime_factorization(_lowerCAmelCase ) lowercase__ : Dict = max(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type int" return ans def a_ ( _lowerCAmelCase : str ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase__ : int = 0 # prime factorization of 'number' lowercase__ : Any = prime_factorization(_lowerCAmelCase ) lowercase__ : str = min(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type int" return ans def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'number' must been an int" assert isinstance(number % 2 == 0 , _lowerCAmelCase ), "compare bust been from type bool" return number % 2 == 0 def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'number' must been an int" assert isinstance(number % 2 != 0 , _lowerCAmelCase ), "compare bust been from type bool" return number % 2 != 0 def a_ ( _lowerCAmelCase : Tuple ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (number > 2) and is_even(_lowerCAmelCase ) ), "'number' must been an int, even and > 2" lowercase__ : Dict = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowercase__ : Optional[Any] = get_prime_numbers(_lowerCAmelCase ) lowercase__ : Optional[int] = len(_lowerCAmelCase ) # run variable for while-loops. lowercase__ : List[Any] = 0 lowercase__ : Any = None # exit variable. for break up the loops lowercase__ : List[Any] = True while i < len_pn and loop: lowercase__ : Dict = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowercase__ : Optional[Any] = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (len(_lowerCAmelCase ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Dict ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowercase__ : List[Any] = 0 while numbera != 0: lowercase__ : Union[str, Any] = numbera % numbera lowercase__ : Any = numbera lowercase__ : Any = rest # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowercase__ : Tuple = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowercase__ : int = prime_factorization(_lowerCAmelCase ) lowercase__ : Dict = prime_factorization(_lowerCAmelCase ) elif numbera == 1 or numbera == 1: lowercase__ : Any = [] lowercase__ : Tuple = [] lowercase__ : Tuple = max(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : str = 0 lowercase__ : List[Any] = 0 lowercase__ : List[Any] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowercase__ : Tuple = prime_fac_a.count(_lowerCAmelCase ) lowercase__ : List[str] = prime_fac_a.count(_lowerCAmelCase ) for _ in range(max(_lowerCAmelCase , _lowerCAmelCase ) ): ans *= n else: lowercase__ : Optional[Any] = prime_fac_a.count(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ): ans *= n done.append(_lowerCAmelCase ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowercase__ : int = prime_fac_a.count(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ): ans *= n done.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'number' must been a positive int" lowercase__ : int = 0 lowercase__ : List[Any] = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(_lowerCAmelCase ): ans += 1 # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and is_prime( _lowerCAmelCase ), "'ans' must been a prime number and from type int" return ans def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' assert ( is_prime(_lowerCAmelCase ) and is_prime(_lowerCAmelCase ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowercase__ : List[str] = p_number_a + 1 # jump to the next number lowercase__ : Union[str, Any] = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(_lowerCAmelCase ): number += 1 while number < p_number_a: ans.append(_lowerCAmelCase ) number += 1 # fetch the next prime number. while not is_prime(_lowerCAmelCase ): number += 1 # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ans[0] != p_number_a and ans[len(_lowerCAmelCase ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def a_ ( _lowerCAmelCase : Optional[int] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 1), "'n' must been int and >= 1" lowercase__ : Any = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(_lowerCAmelCase ) # precondition assert ans[0] == 1 and ans[len(_lowerCAmelCase ) - 1] == n, "Error in function getDivisiors(...)" return ans def a_ ( _lowerCAmelCase : Tuple ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number > 1 ), "'number' must been an int and >= 1" lowercase__ : int = get_divisors(_lowerCAmelCase ) # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (divisors[0] == 1) and (divisors[len(_lowerCAmelCase ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowercase__ : Dict = gcd(abs(_lowerCAmelCase ) , abs(_lowerCAmelCase ) ) # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def a_ ( _lowerCAmelCase : Dict ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'n' must been a int and >= 0" lowercase__ : Optional[Any] = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def a_ ( _lowerCAmelCase : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'n' must been an int and >= 0" lowercase__ : Union[str, Any] = 0 lowercase__ : Optional[Any] = 1 lowercase__ : Union[str, Any] = 1 # this will be return for _ in range(n - 1 ): lowercase__ : Optional[int] = ans ans += fiba lowercase__ : List[Any] = tmp return ans
645
"""simple docstring""" import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class UpperCAmelCase_ ( unittest.TestCase): def __init__( self , a , a=1_3 , a=7 , a=True , a=True , a=True , a=True , a=9_9 , a=3_2 , a=5 , a=4 , a=3_7 , a="gelu" , a=0.1 , a=0.1 , a=5_1_2 , a=1_6 , a=2 , a=0.02 , a=4 , ) -> Dict: lowercase__ : Optional[Any] = parent lowercase__ : Dict = batch_size lowercase__ : List[Any] = seq_length lowercase__ : int = is_training lowercase__ : str = use_attention_mask lowercase__ : Dict = use_token_type_ids lowercase__ : Optional[int] = use_labels lowercase__ : Tuple = vocab_size lowercase__ : List[str] = hidden_size lowercase__ : Union[str, Any] = num_hidden_layers lowercase__ : int = num_attention_heads lowercase__ : Dict = intermediate_size lowercase__ : List[str] = hidden_act lowercase__ : Dict = hidden_dropout_prob lowercase__ : Tuple = attention_probs_dropout_prob lowercase__ : List[str] = max_position_embeddings lowercase__ : int = type_vocab_size lowercase__ : List[str] = type_sequence_label_size lowercase__ : Union[str, Any] = initializer_range lowercase__ : Optional[int] = num_choices def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : str = None if self.use_attention_mask: lowercase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ : List[str] = None if self.use_token_type_ids: lowercase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__ : Any = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[int] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ , lowercase__ : Tuple = config_and_inputs lowercase__ : Union[str, Any] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Tuple = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Union[str, Any] = FlaxAlbertModelTester(self ) @slow def _UpperCAmelCase ( self ) -> str: for model_class_name in self.all_model_classes: lowercase__ : str = model_class_name.from_pretrained('albert-base-v2' ) lowercase__ : Tuple = model(np.ones((1, 1) ) ) self.assertIsNotNone(a ) @require_flax class UpperCAmelCase_ ( unittest.TestCase): @slow def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : str = FlaxAlbertModel.from_pretrained('albert-base-v2' ) lowercase__ : Optional[int] = np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) lowercase__ : Optional[Any] = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) lowercase__ : Any = model(a , attention_mask=a )[0] lowercase__ : Tuple = (1, 1_1, 7_6_8) self.assertEqual(output.shape , a ) lowercase__ : Optional[Any] = np.array( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , a , atol=1e-4 ) )
645
1
'''simple docstring''' def _lowerCAmelCase ( lowerCamelCase_ : Optional[Any] = 1_0 , lowerCamelCase_ : Optional[Any] = 1_0_0_0 , lowerCamelCase_ : str = True ): assert ( isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError('''Invalid value for min_val or max_val (min_value < max_value)''' ) return min_val if option else max_val def _lowerCAmelCase ( lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Dict ): return int((number_a + number_a) / 2 ) def _lowerCAmelCase ( lowerCamelCase_ : List[str] , lowerCamelCase_ : List[Any] , lowerCamelCase_ : List[Any] ): assert ( isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError('''argument value for lower and higher must be(lower > higher)''' ) if not lower < to_guess < higher: raise ValueError( '''guess value must be within the range of lower and higher value''' ) def answer(lowerCamelCase_ : Dict ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print('''started...''' ) __lowercase = lower __lowercase = higher __lowercase = [] while True: __lowercase = get_avg(__snake_case , __snake_case ) last_numbers.append(__snake_case ) if answer(__snake_case ) == "low": __lowercase = number elif answer(__snake_case ) == "high": __lowercase = number else: break print(f"guess the number : {last_numbers[-1]}" ) print(f"details : {last_numbers!s}" ) def _lowerCAmelCase ( ): __lowercase = int(input('''Enter lower value : ''' ).strip() ) __lowercase = int(input('''Enter high value : ''' ).strip() ) __lowercase = int(input('''Enter value to guess : ''' ).strip() ) guess_the_number(__snake_case , __snake_case , __snake_case ) if __name__ == "__main__": main()
502
"""simple docstring""" from ..utils import DummyObject, requires_backends class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'transformers', 'onnx'] def __init__( self: Union[str, Any] , *_SCREAMING_SNAKE_CASE: Any , **_SCREAMING_SNAKE_CASE: int) -> List[Any]: """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Union[str, Any] , *_SCREAMING_SNAKE_CASE: List[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: int , *_SCREAMING_SNAKE_CASE: Any , **_SCREAMING_SNAKE_CASE: str) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'transformers', 'onnx'] def __init__( self: Tuple , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: List[str]) -> Optional[int]: """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Dict , *_SCREAMING_SNAKE_CASE: Optional[Any] , **_SCREAMING_SNAKE_CASE: int) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[Any] , *_SCREAMING_SNAKE_CASE: Dict , **_SCREAMING_SNAKE_CASE: Optional[Any]) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'transformers', 'onnx'] def __init__( self: int , *_SCREAMING_SNAKE_CASE: List[Any] , **_SCREAMING_SNAKE_CASE: int) -> List[str]: """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Dict , *_SCREAMING_SNAKE_CASE: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: List[Any] , *_SCREAMING_SNAKE_CASE: List[str] , **_SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'transformers', 'onnx'] def __init__( self: Tuple , *_SCREAMING_SNAKE_CASE: List[str] , **_SCREAMING_SNAKE_CASE: List[str]) -> Optional[Any]: """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: List[str] , *_SCREAMING_SNAKE_CASE: Tuple , **_SCREAMING_SNAKE_CASE: int) -> int: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[int] , *_SCREAMING_SNAKE_CASE: str , **_SCREAMING_SNAKE_CASE: Any) -> int: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'transformers', 'onnx'] def __init__( self: Optional[Any] , *_SCREAMING_SNAKE_CASE: Union[str, Any] , **_SCREAMING_SNAKE_CASE: Optional[int]) -> str: """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Optional[int] , *_SCREAMING_SNAKE_CASE: List[Any] , **_SCREAMING_SNAKE_CASE: Tuple) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: int , *_SCREAMING_SNAKE_CASE: str , **_SCREAMING_SNAKE_CASE: Any) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) class A__ ( metaclass=__SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['torch', 'transformers', 'onnx'] def __init__( self: List[Any] , *_SCREAMING_SNAKE_CASE: List[str] , **_SCREAMING_SNAKE_CASE: List[str]) -> int: """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: Union[str, Any] , *_SCREAMING_SNAKE_CASE: Optional[int] , **_SCREAMING_SNAKE_CASE: Optional[Any]) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls: str , *_SCREAMING_SNAKE_CASE: Optional[int] , **_SCREAMING_SNAKE_CASE: Dict) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"])
293
0
def SCREAMING_SNAKE_CASE_ ( __A : Optional[Any] ) -> Optional[Any]: """simple docstring""" if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): a_ : Union[str, Any] = F"""Input value of [number={number}] must be an integer""" raise TypeError(_lowerCAmelCase ) if number < 1: a_ : Dict = F"""Input value of [number={number}] must be > 0""" raise ValueError(_lowerCAmelCase ) a_ : Optional[int] = 1 for i in range(1 , _lowerCAmelCase ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
701
def SCREAMING_SNAKE_CASE_ ( __A : int ) -> bool: """simple docstring""" a_ : List[Any] = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
443
0
from ..utils import DummyObject, requires_backends class _lowercase ( metaclass=A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = ['''speech'''] def __init__( self :Union[str, Any] , *lowerCAmelCase__ :Any , **lowerCAmelCase__ :Optional[Any] ) -> Union[str, Any]: requires_backends(self , ['''speech'''] ) class _lowercase ( metaclass=A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = ['''speech'''] def __init__( self :str , *lowerCAmelCase__ :List[str] , **lowerCAmelCase__ :Optional[int] ) -> Dict: requires_backends(self , ['''speech'''] )
696
from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def _UpperCamelCase ( lowercase__ ): __SCREAMING_SNAKE_CASE : Tuple = prime_factors(lowercase__ ) if is_square_free(lowercase__ ): return -1 if len(lowercase__ ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
696
1
import argparse import ast import logging import os import sys import pandas as pd import torch from tqdm import tqdm from transformers import BartForConditionalGeneration, RagRetriever, RagSequenceForGeneration, RagTokenForGeneration from transformers import logging as transformers_logging sys.path.append(os.path.join(os.getcwd())) # noqa: E402 # isort:skip from utils_rag import exact_match_score, fa_score # noqa: E402 # isort:skip a_ : List[str] = logging.getLogger(__name__) logging.basicConfig(level=logging.INFO) transformers_logging.set_verbosity_info() def _SCREAMING_SNAKE_CASE ( snake_case_ : Tuple ): if "token" in model_name_or_path: return "rag_token" if "sequence" in model_name_or_path: return "rag_sequence" if "bart" in model_name_or_path: return "bart" return None def _SCREAMING_SNAKE_CASE ( snake_case_ : Tuple , snake_case_ : Optional[Any] , snake_case_ : Tuple ): return max(metric_fn(snake_case_ , snake_case_ ) for gt in ground_truths ) def _SCREAMING_SNAKE_CASE ( snake_case_ : List[Any] , snake_case_ : Tuple , snake_case_ : str ): __magic_name__ = [line.strip() for line in open(snake_case_ , '''r''' ).readlines()] __magic_name__ = [] if args.gold_data_mode == "qa": __magic_name__ = pd.read_csv(snake_case_ , sep='''\t''' , header=snake_case_ ) for answer_list in data[1]: __magic_name__ = ast.literal_eval(snake_case_ ) answers.append(snake_case_ ) else: __magic_name__ = [line.strip() for line in open(snake_case_ , '''r''' ).readlines()] __magic_name__ = [[reference] for reference in references] __magic_name__ = __magic_name__ = __magic_name__ = 0 for prediction, ground_truths in zip(snake_case_ , snake_case_ ): total += 1 em += metric_max_over_ground_truths(snake_case_ , snake_case_ , snake_case_ ) fa += metric_max_over_ground_truths(snake_case_ , snake_case_ , snake_case_ ) __magic_name__ = 100.0 * em / total __magic_name__ = 100.0 * fa / total logger.info(f'F1: {fa:.2f}' ) logger.info(f'EM: {em:.2f}' ) def _SCREAMING_SNAKE_CASE ( snake_case_ : Dict , snake_case_ : int , snake_case_ : Optional[int] ): __magic_name__ = args.k __magic_name__ = [line.strip() for line in open(snake_case_ , '''r''' ).readlines()] __magic_name__ = [line.strip() for line in open(snake_case_ , '''r''' ).readlines()] __magic_name__ = __magic_name__ = 0 for hypo, reference in zip(snake_case_ , snake_case_ ): __magic_name__ = set(hypo.split('''\t''' )[:k] ) __magic_name__ = set(reference.split('''\t''' ) ) total += 1 em += len(hypo_provenance & ref_provenance ) / k __magic_name__ = 100.0 * em / total logger.info(f'Precision@{k}: {em: .2f}' ) def _SCREAMING_SNAKE_CASE ( snake_case_ : Any , snake_case_ : Dict , snake_case_ : List[str] ): def strip_title(snake_case_ : List[str] ): if title.startswith('''"''' ): __magic_name__ = title[1:] if title.endswith('''"''' ): __magic_name__ = title[:-1] return title __magic_name__ = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( snake_case_ , return_tensors='''pt''' , padding=snake_case_ , truncation=snake_case_ , )['''input_ids'''].to(args.device ) __magic_name__ = rag_model.rag.question_encoder(snake_case_ ) __magic_name__ = question_enc_outputs[0] __magic_name__ = rag_model.retriever( snake_case_ , question_enc_pool_output.cpu().detach().to(torch.floataa ).numpy() , prefix=rag_model.rag.generator.config.prefix , n_docs=rag_model.config.n_docs , return_tensors='''pt''' , ) __magic_name__ = rag_model.retriever.index.get_doc_dicts(result.doc_ids ) __magic_name__ = [] for docs in all_docs: __magic_name__ = [strip_title(snake_case_ ) for title in docs['''title''']] provenance_strings.append('''\t'''.join(snake_case_ ) ) return provenance_strings def _SCREAMING_SNAKE_CASE ( snake_case_ : Union[str, Any] , snake_case_ : Optional[int] , snake_case_ : Dict ): with torch.no_grad(): __magic_name__ = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( snake_case_ , return_tensors='''pt''' , padding=snake_case_ , truncation=snake_case_ ) __magic_name__ = inputs_dict.input_ids.to(args.device ) __magic_name__ = inputs_dict.attention_mask.to(args.device ) __magic_name__ = rag_model.generate( # rag_model overwrites generate snake_case_ , attention_mask=snake_case_ , num_beams=args.num_beams , min_length=args.min_length , max_length=args.max_length , early_stopping=snake_case_ , num_return_sequences=1 , bad_words_ids=[[0, 0]] , ) __magic_name__ = rag_model.retriever.generator_tokenizer.batch_decode(snake_case_ , skip_special_tokens=snake_case_ ) if args.print_predictions: for q, a in zip(snake_case_ , snake_case_ ): logger.info('''Q: {} - A: {}'''.format(snake_case_ , snake_case_ ) ) return answers def _SCREAMING_SNAKE_CASE ( ): __magic_name__ = argparse.ArgumentParser() parser.add_argument( '''--model_type''' , choices=['''rag_sequence''', '''rag_token''', '''bart'''] , type=snake_case_ , help=( '''RAG model type: rag_sequence, rag_token or bart, if none specified, the type is inferred from the''' ''' model_name_or_path''' ) , ) parser.add_argument( '''--index_name''' , default=snake_case_ , choices=['''exact''', '''compressed''', '''legacy'''] , type=snake_case_ , help='''RAG model retriever type''' , ) parser.add_argument( '''--index_path''' , default=snake_case_ , type=snake_case_ , help='''Path to the retrieval index''' , ) parser.add_argument('''--n_docs''' , default=5 , type=snake_case_ , help='''Number of retrieved docs''' ) parser.add_argument( '''--model_name_or_path''' , default=snake_case_ , type=snake_case_ , required=snake_case_ , help='''Path to pretrained checkpoints or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--eval_mode''' , choices=['''e2e''', '''retrieval'''] , default='''e2e''' , type=snake_case_ , help=( '''Evaluation mode, e2e calculates exact match and F1 of the downstream task, retrieval calculates''' ''' precision@k.''' ) , ) parser.add_argument('''--k''' , default=1 , type=snake_case_ , help='''k for the precision@k calculation''' ) parser.add_argument( '''--evaluation_set''' , default=snake_case_ , type=snake_case_ , required=snake_case_ , help='''Path to a file containing evaluation samples''' , ) parser.add_argument( '''--gold_data_path''' , default=snake_case_ , type=snake_case_ , required=snake_case_ , help='''Path to a tab-separated file with gold samples''' , ) parser.add_argument( '''--gold_data_mode''' , default='''qa''' , type=snake_case_ , choices=['''qa''', '''ans'''] , help=( '''Format of the gold data file''' '''qa - a single line in the following format: question [tab] answer_list''' '''ans - a single line of the gold file contains the expected answer string''' ) , ) parser.add_argument( '''--predictions_path''' , type=snake_case_ , default='''predictions.txt''' , help='''Name of the predictions file, to be stored in the checkpoints directory''' , ) parser.add_argument( '''--eval_all_checkpoints''' , action='''store_true''' , help='''Evaluate all checkpoints starting with the same prefix as model_name ending and ending with step number''' , ) parser.add_argument( '''--eval_batch_size''' , default=8 , type=snake_case_ , help='''Batch size per GPU/CPU for evaluation.''' , ) parser.add_argument( '''--recalculate''' , help='''Recalculate predictions even if the prediction file exists''' , action='''store_true''' , ) parser.add_argument( '''--num_beams''' , default=4 , type=snake_case_ , help='''Number of beams to be used when generating answers''' , ) parser.add_argument('''--min_length''' , default=1 , type=snake_case_ , help='''Min length of the generated answers''' ) parser.add_argument('''--max_length''' , default=50 , type=snake_case_ , help='''Max length of the generated answers''' ) parser.add_argument( '''--print_predictions''' , action='''store_true''' , help='''If True, prints predictions while evaluating.''' , ) parser.add_argument( '''--print_docs''' , action='''store_true''' , help='''If True, prints docs retried while generating.''' , ) __magic_name__ = parser.parse_args() __magic_name__ = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) return args def _SCREAMING_SNAKE_CASE ( snake_case_ : str ): __magic_name__ = {} if args.model_type is None: __magic_name__ = infer_model_type(args.model_name_or_path ) assert args.model_type is not None if args.model_type.startswith('''rag''' ): __magic_name__ = RagTokenForGeneration if args.model_type == '''rag_token''' else RagSequenceForGeneration __magic_name__ = args.n_docs if args.index_name is not None: __magic_name__ = args.index_name if args.index_path is not None: __magic_name__ = args.index_path else: __magic_name__ = BartForConditionalGeneration __magic_name__ = ( [f.path for f in os.scandir(args.model_name_or_path ) if f.is_dir()] if args.eval_all_checkpoints else [args.model_name_or_path] ) logger.info('''Evaluate the following checkpoints: %s''' , snake_case_ ) __magic_name__ = get_scores if args.eval_mode == '''e2e''' else get_precision_at_k __magic_name__ = evaluate_batch_eae if args.eval_mode == '''e2e''' else evaluate_batch_retrieval for checkpoint in checkpoints: if os.path.exists(args.predictions_path ) and (not args.recalculate): logger.info('''Calculating metrics based on an existing predictions file: {}'''.format(args.predictions_path ) ) score_fn(snake_case_ , args.predictions_path , args.gold_data_path ) continue logger.info('''***** Running evaluation for {} *****'''.format(snake_case_ ) ) logger.info(''' Batch size = %d''' , args.eval_batch_size ) logger.info(''' Predictions will be stored under {}'''.format(args.predictions_path ) ) if args.model_type.startswith('''rag''' ): __magic_name__ = RagRetriever.from_pretrained(snake_case_ , **snake_case_ ) __magic_name__ = model_class.from_pretrained(snake_case_ , retriever=snake_case_ , **snake_case_ ) model.retriever.init_retrieval() else: __magic_name__ = model_class.from_pretrained(snake_case_ , **snake_case_ ) model.to(args.device ) with open(args.evaluation_set , '''r''' ) as eval_file, open(args.predictions_path , '''w''' ) as preds_file: __magic_name__ = [] for line in tqdm(snake_case_ ): questions.append(line.strip() ) if len(snake_case_ ) == args.eval_batch_size: __magic_name__ = evaluate_batch_fn(snake_case_ , snake_case_ , snake_case_ ) preds_file.write('''\n'''.join(snake_case_ ) + '''\n''' ) preds_file.flush() __magic_name__ = [] if len(snake_case_ ) > 0: __magic_name__ = evaluate_batch_fn(snake_case_ , snake_case_ , snake_case_ ) preds_file.write('''\n'''.join(snake_case_ ) ) preds_file.flush() score_fn(snake_case_ , args.predictions_path , args.gold_data_path ) if __name__ == "__main__": a_ : Tuple = get_args() main(args)
713
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler a_ : Optional[int] = 16 a_ : int = 32 def _SCREAMING_SNAKE_CASE ( snake_case_ : Accelerator , snake_case_ : int = 16 , snake_case_ : str = "bert-base-cased" ): __magic_name__ = AutoTokenizer.from_pretrained(snake_case_ ) __magic_name__ = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(snake_case_ : Union[str, Any] ): # max_length=None => use the model max length (it's actually the default) __magic_name__ = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=snake_case_ , max_length=snake_case_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __magic_name__ = datasets.map( snake_case_ , batched=snake_case_ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=snake_case_ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __magic_name__ = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(snake_case_ : Any ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(snake_case_ , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(snake_case_ , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __magic_name__ = DataLoader( tokenized_datasets['''train'''] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=snake_case_ ) __magic_name__ = DataLoader( tokenized_datasets['''validation'''] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=snake_case_ ) return train_dataloader, eval_dataloader def _SCREAMING_SNAKE_CASE ( snake_case_ : str , snake_case_ : Dict , snake_case_ : List[Any] , snake_case_ : str ): model.eval() __magic_name__ = 0 for step, batch in enumerate(snake_case_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __magic_name__ = model(**snake_case_ ) __magic_name__ = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times __magic_name__ , __magic_name__ = accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(snake_case_ ) - 1: __magic_name__ = predictions[: len(eval_dataloader.dataset ) - samples_seen] __magic_name__ = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=snake_case_ , references=snake_case_ , ) __magic_name__ = metric.compute() return eval_metric["accuracy"] def _SCREAMING_SNAKE_CASE ( snake_case_ : Union[str, Any] , snake_case_ : Tuple ): # Initialize accelerator __magic_name__ = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __magic_name__ = config['''lr'''] __magic_name__ = int(config['''num_epochs'''] ) __magic_name__ = int(config['''seed'''] ) __magic_name__ = int(config['''batch_size'''] ) __magic_name__ = args.model_name_or_path set_seed(snake_case_ ) __magic_name__ , __magic_name__ = get_dataloaders(snake_case_ , snake_case_ , snake_case_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __magic_name__ = AutoModelForSequenceClassification.from_pretrained(snake_case_ , return_dict=snake_case_ ) # Instantiate optimizer __magic_name__ = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __magic_name__ = optimizer_cls(params=model.parameters() , lr=snake_case_ ) if accelerator.state.deepspeed_plugin is not None: __magic_name__ = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __magic_name__ = 1 __magic_name__ = (len(snake_case_ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __magic_name__ = get_linear_schedule_with_warmup( optimizer=snake_case_ , num_warmup_steps=0 , num_training_steps=snake_case_ , ) else: __magic_name__ = DummyScheduler(snake_case_ , total_num_steps=snake_case_ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # We need to keep track of how many total steps we have iterated over __magic_name__ = 0 # We also need to keep track of the stating epoch so files are named properly __magic_name__ = 0 __magic_name__ = evaluate.load('''glue''' , '''mrpc''' ) __magic_name__ = num_epochs if args.partial_train_epoch is not None: __magic_name__ = args.partial_train_epoch if args.resume_from_checkpoint: accelerator.load_state(args.resume_from_checkpoint ) __magic_name__ = args.resume_from_checkpoint.split('''epoch_''' )[1] __magic_name__ = '''''' for char in epoch_string: if char.isdigit(): state_epoch_num += char else: break __magic_name__ = int(snake_case_ ) + 1 __magic_name__ = evaluation_loop(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) accelerator.print('''resumed checkpoint performance:''' , snake_case_ ) accelerator.print('''resumed checkpoint\'s scheduler\'s lr:''' , lr_scheduler.get_lr()[0] ) accelerator.print('''resumed optimizers\'s lr:''' , optimizer.param_groups[0]['''lr'''] ) with open(os.path.join(args.output_dir , f'state_{starting_epoch-1}.json' ) , '''r''' ) as f: __magic_name__ = json.load(snake_case_ ) assert resumed_state["accuracy"] == accuracy, "Accuracy mismatch, loading from checkpoint failed" assert ( resumed_state["lr"] == lr_scheduler.get_lr()[0] ), "Scheduler learning rate mismatch, loading from checkpoint failed" assert ( resumed_state["optimizer_lr"] == optimizer.param_groups[0]["lr"] ), "Optimizer learning rate mismatch, loading from checkpoint failed" assert resumed_state["epoch"] == starting_epoch - 1, "Epoch mismatch, loading from checkpoint failed" return # Now we train the model __magic_name__ = {} for epoch in range(snake_case_ , snake_case_ ): model.train() for step, batch in enumerate(snake_case_ ): __magic_name__ = model(**snake_case_ ) __magic_name__ = outputs.loss __magic_name__ = loss / gradient_accumulation_steps accelerator.backward(snake_case_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 __magic_name__ = f'epoch_{epoch}' __magic_name__ = os.path.join(args.output_dir , snake_case_ ) accelerator.save_state(snake_case_ ) __magic_name__ = evaluation_loop(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) __magic_name__ = accuracy __magic_name__ = lr_scheduler.get_lr()[0] __magic_name__ = optimizer.param_groups[0]['''lr'''] __magic_name__ = epoch __magic_name__ = overall_step accelerator.print(f'epoch {epoch}:' , snake_case_ ) accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , f'state_{epoch}.json' ) , '''w''' ) as f: json.dump(snake_case_ , snake_case_ ) def _SCREAMING_SNAKE_CASE ( ): __magic_name__ = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=snake_case_ , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=snake_case_ , ) parser.add_argument( '''--output_dir''' , type=snake_case_ , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--resume_from_checkpoint''' , type=snake_case_ , default=snake_case_ , help='''If the training should continue from a checkpoint folder.''' , ) parser.add_argument( '''--partial_train_epoch''' , type=snake_case_ , default=snake_case_ , help='''If passed, the training will stop after this number of epochs.''' , ) parser.add_argument( '''--num_epochs''' , type=snake_case_ , default=2 , help='''Number of train epochs.''' , ) __magic_name__ = parser.parse_args() __magic_name__ = {'''lr''': 2E-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(snake_case_ , snake_case_ ) if __name__ == "__main__": main()
678
0
import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer __a = logging.get_logger(__name__) __a = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} __a = { 'vocab_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } __a = { 'vocab_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } __a = { 'vocab_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json' ), }, } __a = { 'facebook/dpr-ctx_encoder-single-nq-base': 512, 'facebook/dpr-ctx_encoder-multiset-base': 512, } __a = { 'facebook/dpr-question_encoder-single-nq-base': 512, 'facebook/dpr-question_encoder-multiset-base': 512, } __a = { 'facebook/dpr-reader-single-nq-base': 512, 'facebook/dpr-reader-multiset-base': 512, } __a = { 'facebook/dpr-ctx_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-ctx_encoder-multiset-base': {'do_lower_case': True}, } __a = { 'facebook/dpr-question_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-question_encoder-multiset-base': {'do_lower_case': True}, } __a = { 'facebook/dpr-reader-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-reader-multiset-base': {'do_lower_case': True}, } class __a( _a ): """simple docstring""" lowerCAmelCase = VOCAB_FILES_NAMES lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION class __a( _a ): """simple docstring""" lowerCAmelCase = VOCAB_FILES_NAMES lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION __a = collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) __a = collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) __a = R'\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n ```\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n ```\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `\'tf\'`: Return TensorFlow `tf.constant` objects.\n - `\'pt\'`: Return PyTorch `torch.Tensor` objects.\n - `\'np\'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer\'s default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Returns:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n ' @add_start_docstrings(_a ) class __a: """simple docstring""" def __call__( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = False ,_SCREAMING_SNAKE_CASE = False ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = None ,**_SCREAMING_SNAKE_CASE ,) -> BatchEncoding: if titles is None and texts is None: return super().__call__( _SCREAMING_SNAKE_CASE ,padding=_SCREAMING_SNAKE_CASE ,truncation=_SCREAMING_SNAKE_CASE ,max_length=_SCREAMING_SNAKE_CASE ,return_tensors=_SCREAMING_SNAKE_CASE ,return_attention_mask=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ,) elif titles is None or texts is None: UpperCAmelCase_ : List[str] = titles if texts is None else texts return super().__call__( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,padding=_SCREAMING_SNAKE_CASE ,truncation=_SCREAMING_SNAKE_CASE ,max_length=_SCREAMING_SNAKE_CASE ,return_tensors=_SCREAMING_SNAKE_CASE ,return_attention_mask=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ,) UpperCAmelCase_ : List[Any] = titles if not isinstance(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) else [titles] UpperCAmelCase_ : List[str] = texts if not isinstance(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) else [texts] UpperCAmelCase_ : Any = len(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = questions if not isinstance(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) else [questions] * n_passages if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( f'''There should be as many titles than texts but got {len(_SCREAMING_SNAKE_CASE )} titles and {len(_SCREAMING_SNAKE_CASE )} texts.''' ) UpperCAmelCase_ : Tuple = super().__call__(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,padding=_SCREAMING_SNAKE_CASE ,truncation=_SCREAMING_SNAKE_CASE )['''input_ids'''] UpperCAmelCase_ : int = super().__call__(_SCREAMING_SNAKE_CASE ,add_special_tokens=_SCREAMING_SNAKE_CASE ,padding=_SCREAMING_SNAKE_CASE ,truncation=_SCREAMING_SNAKE_CASE )['''input_ids'''] UpperCAmelCase_ : Optional[int] = { '''input_ids''': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) ] } if return_attention_mask is not False: UpperCAmelCase_ : List[str] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) UpperCAmelCase_ : Dict = attention_mask return self.pad(_SCREAMING_SNAKE_CASE ,padding=_SCREAMING_SNAKE_CASE ,max_length=_SCREAMING_SNAKE_CASE ,return_tensors=_SCREAMING_SNAKE_CASE ) def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = 16 ,_SCREAMING_SNAKE_CASE = 64 ,_SCREAMING_SNAKE_CASE = 4 ,) -> List[DPRSpanPrediction]: UpperCAmelCase_ : Tuple = reader_input['''input_ids'''] UpperCAmelCase_, UpperCAmelCase_, UpperCAmelCase_ : Optional[Any] = reader_output[:3] UpperCAmelCase_ : Optional[Any] = len(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = sorted(range(_SCREAMING_SNAKE_CASE ) ,reverse=_SCREAMING_SNAKE_CASE ,key=relevance_logits.__getitem__ ) UpperCAmelCase_ : List[DPRReaderOutput] = [] for doc_id in sorted_docs: UpperCAmelCase_ : List[Any] = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence UpperCAmelCase_ : str = sequence_ids.index(self.sep_token_id ,2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: UpperCAmelCase_ : List[Any] = sequence_ids.index(self.pad_token_id ) else: UpperCAmelCase_ : int = len(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Tuple = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] ,end_logits=end_logits[doc_id][passage_offset:sequence_len] ,max_answer_length=_SCREAMING_SNAKE_CASE ,top_spans=_SCREAMING_SNAKE_CASE ,) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] ,relevance_score=relevance_logits[doc_id] ,doc_id=_SCREAMING_SNAKE_CASE ,start_index=_SCREAMING_SNAKE_CASE ,end_index=_SCREAMING_SNAKE_CASE ,text=self.decode(sequence_ids[start_index : end_index + 1] ) ,) ) if len(_SCREAMING_SNAKE_CASE ) >= num_spans: break return nbest_spans_predictions[:num_spans] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,) -> List[DPRSpanPrediction]: UpperCAmelCase_ : Tuple = [] for start_index, start_score in enumerate(_SCREAMING_SNAKE_CASE ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) UpperCAmelCase_ : int = sorted(_SCREAMING_SNAKE_CASE ,key=lambda _SCREAMING_SNAKE_CASE : x[1] ,reverse=_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Union[str, Any] = [] for (start_index, end_index), score in scores: if start_index > end_index: raise ValueError(f'''Wrong span indices: [{start_index}:{end_index}]''' ) UpperCAmelCase_ : str = end_index - start_index + 1 if length > max_answer_length: raise ValueError(f'''Span is too long: {length} > {max_answer_length}''' ) if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(_SCREAMING_SNAKE_CASE ) == top_spans: break return chosen_span_intervals @add_end_docstrings(_a ) class __a( _a , _a ): """simple docstring""" lowerCAmelCase = VOCAB_FILES_NAMES lowerCAmelCase = READER_PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase = READER_PRETRAINED_INIT_CONFIGURATION lowerCAmelCase = ['''input_ids''', '''attention_mask''']
30
"""simple docstring""" from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class a__ ( A__ ): UpperCAmelCase__ = '''''' UpperCAmelCase__ = '''hf-legacy''' # "hf://"" is reserved for hffs def __init__( self :Dict , _lowerCamelCase :Optional[DatasetInfo] = None , _lowerCamelCase :Optional[str] = None , **_lowerCamelCase :Tuple , ): '''simple docstring''' super().__init__(self , **_lowerCamelCase ) UpperCamelCase_ : List[str] =repo_info UpperCamelCase_ : Any =token UpperCamelCase_ : Tuple =None def lowerCamelCase_ ( self :Dict ): '''simple docstring''' if self.dir_cache is None: UpperCamelCase_ : Any ={} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes UpperCamelCase_ : Optional[Any] ={ 'name': hf_file.rfilename, 'size': None, 'type': 'file', } self.dir_cache.update( { str(_lowerCamelCase ): {'name': str(_lowerCamelCase ), 'size': None, 'type': 'directory'} for d in list(PurePosixPath(hf_file.rfilename ).parents )[:-1] } ) def lowerCamelCase_ ( self :Union[str, Any] , _lowerCamelCase :str , _lowerCamelCase :str = "rb" , **_lowerCamelCase :str , ): '''simple docstring''' if not isinstance(self.repo_info , _lowerCamelCase ): raise NotImplementedError(f'''Open is only implemented for dataset repositories, but got {self.repo_info}''' ) UpperCamelCase_ : List[Any] =hf_hub_url(self.repo_info.id , _lowerCamelCase , revision=self.repo_info.sha ) return fsspec.open( _lowerCamelCase , mode=_lowerCamelCase , headers=get_authentication_headers_for_url(_lowerCamelCase , use_auth_token=self.token ) , client_kwargs={'trust_env': True} , ).open() def lowerCamelCase_ ( self :Optional[Any] , _lowerCamelCase :Tuple , **_lowerCamelCase :Any ): '''simple docstring''' self._get_dirs() UpperCamelCase_ : Tuple =self._strip_protocol(_lowerCamelCase ) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(_lowerCamelCase ) def lowerCamelCase_ ( self :Optional[Any] , _lowerCamelCase :List[str] , _lowerCamelCase :List[Any]=False , **_lowerCamelCase :Any ): '''simple docstring''' self._get_dirs() UpperCamelCase_ : str =PurePosixPath(path.strip('/' ) ) UpperCamelCase_ : List[str] ={} for p, f in self.dir_cache.items(): UpperCamelCase_ : List[Any] =PurePosixPath(p.strip('/' ) ) UpperCamelCase_ : Tuple =p.parent if root == path: UpperCamelCase_ : int =f UpperCamelCase_ : Optional[int] =list(paths.values() ) if detail: return out else: return sorted(f['name'] for f in out )
357
0
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase :str = logging.get_logger(__name__) __lowerCamelCase :Any = { 'huggingface/time-series-transformer-tourism-monthly': ( 'https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json' ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class A__ ( __lowercase): """simple docstring""" snake_case__ : List[Any] ='''time_series_transformer''' snake_case__ : List[Any] ={ '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self: List[str] , __a: Optional[int] = None , __a: Optional[int] = None , __a: str = "student_t" , __a: str = "nll" , __a: int = 1 , __a: List[int] = [1, 2, 3, 4, 5, 6, 7] , __a: Optional[Union[str, bool]] = "mean" , __a: int = 0 , __a: int = 0 , __a: int = 0 , __a: int = 0 , __a: Optional[List[int]] = None , __a: Optional[List[int]] = None , __a: int = 32 , __a: int = 32 , __a: int = 2 , __a: int = 2 , __a: int = 2 , __a: int = 2 , __a: bool = True , __a: str = "gelu" , __a: int = 64 , __a: float = 0.1 , __a: float = 0.1 , __a: float = 0.1 , __a: float = 0.1 , __a: float = 0.1 , __a: int = 100 , __a: float = 0.02 , __a: Tuple=True , **__a: str , )-> Any: # time series specific configuration lowerCamelCase : str = prediction_length lowerCamelCase : Optional[Any] = context_length or prediction_length lowerCamelCase : Tuple = distribution_output lowerCamelCase : Any = loss lowerCamelCase : List[Any] = input_size lowerCamelCase : int = num_time_features lowerCamelCase : Dict = lags_sequence lowerCamelCase : Optional[int] = scaling lowerCamelCase : int = num_dynamic_real_features lowerCamelCase : Tuple = num_static_real_features lowerCamelCase : Any = num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(__a ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) lowerCamelCase : int = cardinality else: lowerCamelCase : Dict = [0] if embedding_dimension and num_static_categorical_features > 0: if len(__a ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) lowerCamelCase : str = embedding_dimension else: lowerCamelCase : str = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] lowerCamelCase : Any = num_parallel_samples # Transformer architecture configuration lowerCamelCase : Any = input_size * len(__a ) + self._number_of_features lowerCamelCase : List[str] = d_model lowerCamelCase : Tuple = encoder_attention_heads lowerCamelCase : Optional[int] = decoder_attention_heads lowerCamelCase : Union[str, Any] = encoder_ffn_dim lowerCamelCase : str = decoder_ffn_dim lowerCamelCase : str = encoder_layers lowerCamelCase : Any = decoder_layers lowerCamelCase : Optional[int] = dropout lowerCamelCase : List[str] = attention_dropout lowerCamelCase : Tuple = activation_dropout lowerCamelCase : Optional[int] = encoder_layerdrop lowerCamelCase : int = decoder_layerdrop lowerCamelCase : Optional[int] = activation_function lowerCamelCase : Optional[Any] = init_std lowerCamelCase : Optional[Any] = use_cache super().__init__(is_encoder_decoder=__a , **__a ) @property def a__ ( self: int )-> int: return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
42
"""simple docstring""" import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class A__ ( __lowercase): """simple docstring""" snake_case__ : Tuple =(KDPMaDiscreteScheduler,) snake_case__ : Tuple =10 def a__ ( self: List[Any] , **__a: Optional[int] )-> Union[str, Any]: lowerCamelCase : int = { """num_train_timesteps""": 1_100, """beta_start""": 0.00_01, """beta_end""": 0.02, """beta_schedule""": """linear""", } config.update(**__a ) return config def a__ ( self: Union[str, Any] )-> Any: for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=__a ) def a__ ( self: str )-> int: for beta_start, beta_end in zip([0.0_00_01, 0.00_01, 0.0_01] , [0.00_02, 0.0_02, 0.02] ): self.check_over_configs(beta_start=__a , beta_end=__a ) def a__ ( self: int )-> Union[str, Any]: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=__a ) def a__ ( self: List[Any] )-> List[Any]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__a ) def a__ ( self: Union[str, Any] )-> int: lowerCamelCase : List[str] = self.scheduler_classes[0] lowerCamelCase : Union[str, Any] = self.get_scheduler_config(prediction_type="""v_prediction""" ) lowerCamelCase : List[str] = scheduler_class(**__a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCamelCase : Dict = self.dummy_model() lowerCamelCase : List[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCamelCase : List[Any] = sample.to(__a ) for i, t in enumerate(scheduler.timesteps ): lowerCamelCase : Optional[Any] = scheduler.scale_model_input(__a , __a ) lowerCamelCase : Optional[int] = model(__a , __a ) lowerCamelCase : Tuple = scheduler.step(__a , __a , __a ) lowerCamelCase : Optional[Any] = output.prev_sample lowerCamelCase : List[str] = torch.sum(torch.abs(__a ) ) lowerCamelCase : Tuple = torch.mean(torch.abs(__a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6_9_3_4e-0_7 ) < 1e-2 assert abs(result_mean.item() - 6.1_1_1_2e-1_0 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 4.6_9_3_4_2_8_6_5_0_1_7_0_9_7_2e-0_7 ) < 1e-2 assert abs(result_mean.item() - 0.00_02 ) < 1e-3 def a__ ( self: Any )-> Any: if torch_device == "mps": return lowerCamelCase : Dict = self.scheduler_classes[0] lowerCamelCase : Dict = self.get_scheduler_config() lowerCamelCase : int = scheduler_class(**__a ) scheduler.set_timesteps(self.num_inference_steps ) lowerCamelCase : List[Any] = self.dummy_model() lowerCamelCase : Optional[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCamelCase : Optional[int] = sample.to(__a ) for i, t in enumerate(scheduler.timesteps ): lowerCamelCase : Dict = scheduler.scale_model_input(__a , __a ) lowerCamelCase : Optional[Any] = model(__a , __a ) lowerCamelCase : Tuple = scheduler.step(__a , __a , __a ) lowerCamelCase : str = output.prev_sample lowerCamelCase : Tuple = torch.sum(torch.abs(__a ) ) lowerCamelCase : Tuple = torch.mean(torch.abs(__a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.41_25 ) < 1e-2 assert abs(result_mean.item() - 0.02_66 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1e-2 assert abs(result_mean.item() - 0.02_66 ) < 1e-3 def a__ ( self: Optional[Any] )-> List[Any]: if torch_device == "mps": return lowerCamelCase : Any = self.scheduler_classes[0] lowerCamelCase : Union[str, Any] = self.get_scheduler_config() lowerCamelCase : Optional[Any] = scheduler_class(**__a ) scheduler.set_timesteps(self.num_inference_steps , device=__a ) lowerCamelCase : Union[str, Any] = self.dummy_model() lowerCamelCase : List[str] = self.dummy_sample_deter.to(__a ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowerCamelCase : Union[str, Any] = scheduler.scale_model_input(__a , __a ) lowerCamelCase : Optional[int] = model(__a , __a ) lowerCamelCase : int = scheduler.step(__a , __a , __a ) lowerCamelCase : int = output.prev_sample lowerCamelCase : Union[str, Any] = torch.sum(torch.abs(__a ) ) lowerCamelCase : int = torch.mean(torch.abs(__a ) ) if str(__a ).startswith("""cpu""" ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.41_25 ) < 1e-2 assert abs(result_mean.item() - 0.02_66 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1e-2 assert abs(result_mean.item() - 0.02_66 ) < 1e-3
42
1
import string def UpperCAmelCase ( a_ ) -> None: """simple docstring""" for key in range(len(string.ascii_uppercase ) ): __A = "" for symbol in message: if symbol in string.ascii_uppercase: __A = string.ascii_uppercase.find(a_ ) __A = num - key if num < 0: __A = num + len(string.ascii_uppercase ) __A = translated + string.ascii_uppercase[num] else: __A = translated + symbol print(F'''Decryption using Key #{key}: {translated}''' ) def UpperCAmelCase ( ) -> None: """simple docstring""" __A = input("Encrypted message: " ) __A = message.upper() decrypt(a_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
55
"""simple docstring""" def lowerCamelCase ( _UpperCamelCase : int = 1_0_0_0_0_0_0 ) -> int: '''simple docstring''' __UpperCAmelCase : List[str] = 1 __UpperCAmelCase : Union[str, Any] = 1 __UpperCAmelCase : Optional[Any] = {1: 1} for inputa in range(2 , _UpperCamelCase ): __UpperCAmelCase : Optional[int] = 0 __UpperCAmelCase : str = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __UpperCAmelCase : Tuple = (3 * number) + 1 counter += 1 if inputa not in counters: __UpperCAmelCase : Optional[Any] = counter if counter > pre_counter: __UpperCAmelCase : List[Any] = inputa __UpperCAmelCase : List[Any] = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
139
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case = { """configuration_biogpt""": ["""BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BioGptConfig"""], """tokenization_biogpt""": ["""BioGptTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BioGptForCausalLM""", """BioGptForTokenClassification""", """BioGptForSequenceClassification""", """BioGptModel""", """BioGptPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
701
import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class lowerCAmelCase : def __init__( self :List[str] , _lowercase :Optional[Any] , _lowercase :str=13 , _lowercase :Tuple=7 , _lowercase :Any=True , _lowercase :Optional[int]=True , _lowercase :Optional[Any]=True , _lowercase :Optional[int]=True , _lowercase :str=99 , _lowercase :Optional[int]=64 , _lowercase :Optional[int]=32 , _lowercase :Union[str, Any]=5 , _lowercase :Optional[int]=4 , _lowercase :Any=37 , _lowercase :Optional[int]="gelu" , _lowercase :Optional[int]=0.1 , _lowercase :str=0.1 , _lowercase :Union[str, Any]=5_12 , _lowercase :Optional[int]=16 , _lowercase :int=2 , _lowercase :Tuple=0.02 , _lowercase :Optional[Any]=3 , _lowercase :Dict=4 , _lowercase :List[Any]=None , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = seq_length lowercase__ = is_training lowercase__ = use_input_mask lowercase__ = use_token_type_ids lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = embedding_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = type_sequence_label_size lowercase__ = initializer_range lowercase__ = num_labels lowercase__ = num_choices lowercase__ = scope def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ = None if self.use_input_mask: lowercase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ = None if self.use_token_type_ids: lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase__ = None lowercase__ = None lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase__ = ids_tensor([self.batch_size] , self.num_choices ) lowercase__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' return MegatronBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowercase , initializer_range=self.initializer_range , ) def UpperCAmelCase ( self :str , _lowercase :Tuple , _lowercase :Tuple , _lowercase :Tuple , _lowercase :Dict , _lowercase :Optional[Any] , _lowercase :int , _lowercase :List[Any] ): '''simple docstring''' lowercase__ = MegatronBertModel(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase ) lowercase__ = model(_lowercase , token_type_ids=_lowercase ) lowercase__ = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCAmelCase ( self :Any , _lowercase :Dict , _lowercase :Union[str, Any] , _lowercase :Optional[int] , _lowercase :Any , _lowercase :List[str] , _lowercase :Any , _lowercase :int ): '''simple docstring''' lowercase__ = MegatronBertForMaskedLM(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase ( self :Dict , _lowercase :str , _lowercase :Optional[Any] , _lowercase :Optional[Any] , _lowercase :List[Any] , _lowercase :Union[str, Any] , _lowercase :Optional[int] , _lowercase :List[str] ): '''simple docstring''' lowercase__ = MegatronBertForCausalLM(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase ( self :Any , _lowercase :int , _lowercase :Tuple , _lowercase :Optional[int] , _lowercase :Dict , _lowercase :Dict , _lowercase :Optional[int] , _lowercase :Dict ): '''simple docstring''' lowercase__ = MegatronBertForNextSentencePrediction(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def UpperCAmelCase ( self :str , _lowercase :int , _lowercase :Union[str, Any] , _lowercase :str , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :Dict , _lowercase :List[str] ): '''simple docstring''' lowercase__ = MegatronBertForPreTraining(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase , next_sentence_label=_lowercase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def UpperCAmelCase ( self :str , _lowercase :Optional[Any] , _lowercase :Tuple , _lowercase :int , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :Any , _lowercase :Any ): '''simple docstring''' lowercase__ = MegatronBertForQuestionAnswering(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , start_positions=_lowercase , end_positions=_lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase ( self :str , _lowercase :str , _lowercase :Any , _lowercase :Dict , _lowercase :Optional[Any] , _lowercase :int , _lowercase :int , _lowercase :Optional[Any] ): '''simple docstring''' lowercase__ = self.num_labels lowercase__ = MegatronBertForSequenceClassification(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase ( self :List[Any] , _lowercase :List[str] , _lowercase :List[str] , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :Optional[Any] ): '''simple docstring''' lowercase__ = self.num_labels lowercase__ = MegatronBertForTokenClassification(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :Union[str, Any] , _lowercase :List[str] , _lowercase :int , _lowercase :int , _lowercase :List[Any] , _lowercase :List[Any] , _lowercase :Tuple ): '''simple docstring''' lowercase__ = self.num_choices lowercase__ = MegatronBertForMultipleChoice(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowercase__ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowercase__ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowercase__ = model( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , labels=_lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() ( ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ( lowercase__ ) , ) = config_and_inputs lowercase__ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowerCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): __lowerCamelCase = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) __lowerCamelCase = ( { 'feature-extraction': MegatronBertModel, 'fill-mask': MegatronBertForMaskedLM, 'question-answering': MegatronBertForQuestionAnswering, 'text-classification': MegatronBertForSequenceClassification, 'text-generation': MegatronBertForCausalLM, 'token-classification': MegatronBertForTokenClassification, 'zero-shot': MegatronBertForSequenceClassification, } if is_torch_available() else {} ) __lowerCamelCase = True # test_resize_embeddings = False __lowerCamelCase = False def UpperCAmelCase ( self :str , _lowercase :Tuple , _lowercase :str , _lowercase :int=False ): '''simple docstring''' lowercase__ = super()._prepare_for_class(_lowercase , _lowercase , return_labels=_lowercase ) if return_labels: if model_class in get_values(_lowercase ): lowercase__ = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=_lowercase ) lowercase__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowercase ) return inputs_dict def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = MegatronBertModelTester(self ) lowercase__ = ConfigTester(self , config_class=_lowercase , hidden_size=37 ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*_lowercase ) def UpperCAmelCase ( self :Any ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*_lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*_lowercase ) def UpperCAmelCase ( self :str ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*_lowercase ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*_lowercase ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*_lowercase ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*_lowercase ) def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*_lowercase ) def _A ( __magic_name__ ): return torch.tensor( __magic_name__ , dtype=torch.long , device=__magic_name__ , ) _snake_case = 1E-4 @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): @slow @unittest.skip("Model is not available." ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = "nvidia/megatron-bert-uncased-345m" if "MYDIR" in os.environ: lowercase__ = os.path.join(os.environ["MYDIR"] , _lowercase ) lowercase__ = MegatronBertModel.from_pretrained(_lowercase ) model.to(_lowercase ) model.half() lowercase__ = _long_tensor([[1_01, 71_10, 10_05, 10_56, 20_23, 1_13_33, 1_74_13, 10_29, 1_02]] ) with torch.no_grad(): lowercase__ = model(_lowercase )[0] lowercase__ = torch.Size((1, 9, 10_24) ) self.assertEqual(output.shape , _lowercase ) lowercase__ = [-0.6040, -0.2517, -0.1025, 0.3420, -0.6758, -0.0017, -0.1089, -0.1990, 0.5728] for ii in range(3 ): for jj in range(3 ): lowercase__ = output[0, ii, jj] lowercase__ = expected[3 * ii + jj] lowercase__ = "ii={} jj={} a={} b={}".format(_lowercase , _lowercase , _lowercase , _lowercase ) self.assertTrue(math.isclose(_lowercase , _lowercase , rel_tol=_lowercase , abs_tol=_lowercase ) , msg=_lowercase )
611
0
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel __lowerCamelCase : Any = False __lowerCamelCase : int = True __lowerCamelCase : int = False if __name__ == "__main__": __lowerCamelCase : Dict = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') __lowerCamelCase : Dict = parser.parse_args() __lowerCamelCase : List[str] = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } __lowerCamelCase : Optional[int] = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } __lowerCamelCase : Union[str, Any] = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: __lowerCamelCase : Any = reader.read() __lowerCamelCase : Dict = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): __lowerCamelCase : List[Any] = UNetaDModel(**config) else: __lowerCamelCase : int = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel __lowerCamelCase : int = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) __lowerCamelCase : List[Any] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: __lowerCamelCase : List[str] = config[key] del config[key] __lowerCamelCase : List[str] = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] __lowerCamelCase : List[Any] = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: __lowerCamelCase : Any = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) __lowerCamelCase : List[Any] = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue __lowerCamelCase : str = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: __lowerCamelCase : Optional[int] = param_value __lowerCamelCase : str = True if not has_changed: __lowerCamelCase : Union[str, Any] = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
653
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase : str = logging.get_logger(__name__) __lowerCamelCase : Tuple = { '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', } class A_ (a_ ): """simple docstring""" a__ = '''gpt_bigcode''' a__ = ['''past_key_values'''] a__ = { '''hidden_size''': '''n_embd''', '''max_position_embeddings''': '''n_positions''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self :List[Any] , lowerCAmelCase__ :Any=50_257 , lowerCAmelCase__ :Dict=1_024 , lowerCAmelCase__ :Optional[int]=768 , lowerCAmelCase__ :Tuple=12 , lowerCAmelCase__ :int=12 , lowerCAmelCase__ :Optional[Any]=None , lowerCAmelCase__ :List[str]="gelu_pytorch_tanh" , lowerCAmelCase__ :Tuple=0.1 , lowerCAmelCase__ :Tuple=0.1 , lowerCAmelCase__ :str=0.1 , lowerCAmelCase__ :Any=1E-5 , lowerCAmelCase__ :Union[str, Any]=0.0_2 , lowerCAmelCase__ :Union[str, Any]=True , lowerCAmelCase__ :Optional[Any]=True , lowerCAmelCase__ :int=50_256 , lowerCAmelCase__ :List[str]=50_256 , lowerCAmelCase__ :List[Any]=True , lowerCAmelCase__ :str=True , lowerCAmelCase__ :int=True , **lowerCAmelCase__ :Union[str, Any] , ) -> Any: '''simple docstring''' snake_case_ : List[Any] = vocab_size snake_case_ : Any = n_positions snake_case_ : Any = n_embd snake_case_ : Optional[Any] = n_layer snake_case_ : List[Any] = n_head snake_case_ : Tuple = n_inner snake_case_ : str = activation_function snake_case_ : Union[str, Any] = resid_pdrop snake_case_ : Optional[Any] = embd_pdrop snake_case_ : Any = attn_pdrop snake_case_ : List[Any] = layer_norm_epsilon snake_case_ : Tuple = initializer_range snake_case_ : int = scale_attn_weights snake_case_ : Union[str, Any] = use_cache snake_case_ : Dict = attention_softmax_in_fpaa snake_case_ : Any = scale_attention_softmax_in_fpaa snake_case_ : List[str] = multi_query snake_case_ : List[str] = bos_token_id snake_case_ : Any = eos_token_id super().__init__(bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ )
653
1
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowercase = logging.get_logger(__name__) lowercase = Dict[str, Any] lowercase = List[Prediction] @add_end_docstrings(_lowercase) class SCREAMING_SNAKE_CASE_ ( _lowercase): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__) -> Optional[int]: '''simple docstring''' super().__init__(*lowerCamelCase__ , **lowerCamelCase__) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""") requires_backends(self , "vision") self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items())) def UpperCAmelCase ( self , **lowerCamelCase__) -> str: '''simple docstring''' snake_case__ : List[Any] = {} if "threshold" in kwargs: snake_case__ : str = kwargs["threshold"] return {}, {}, postprocess_kwargs def __call__( self , *lowerCamelCase__ , **lowerCamelCase__) -> Union[Predictions, List[Prediction]]: '''simple docstring''' return super().__call__(*lowerCamelCase__ , **lowerCamelCase__) def UpperCAmelCase ( self , lowerCamelCase__) -> Union[str, Any]: '''simple docstring''' snake_case__ : List[str] = load_image(lowerCamelCase__) snake_case__ : int = torch.IntTensor([[image.height, image.width]]) snake_case__ : Dict = self.image_processor(images=[image] , return_tensors="pt") if self.tokenizer is not None: snake_case__ : List[Any] = self.tokenizer(text=inputs["words"] , boxes=inputs["boxes"] , return_tensors="pt") snake_case__ : List[str] = target_size return inputs def UpperCAmelCase ( self , lowerCamelCase__) -> Optional[int]: '''simple docstring''' snake_case__ : Optional[int] = model_inputs.pop("target_size") snake_case__ : Optional[int] = self.model(**lowerCamelCase__) snake_case__ : List[str] = outputs.__class__({"target_size": target_size, **outputs}) if self.tokenizer is not None: snake_case__ : str = model_inputs["bbox"] return model_outputs def UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__=0.9) -> Optional[Any]: '''simple docstring''' snake_case__ : Dict = model_outputs["target_size"] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. snake_case__, snake_case__ : Optional[int] = target_size[0].tolist() def unnormalize(lowerCamelCase__): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1_000), (height * bbox[1] / 1_000), (width * bbox[2] / 1_000), (height * bbox[3] / 1_000), ])) snake_case__, snake_case__ : Optional[Any] = model_outputs["logits"].squeeze(0).softmax(dim=-1).max(dim=-1) snake_case__ : List[str] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] snake_case__ : Union[str, Any] = [unnormalize(lowerCamelCase__) for bbox in model_outputs["bbox"].squeeze(0)] snake_case__ : Optional[int] = ["score", "label", "box"] snake_case__ : Union[str, Any] = [dict(zip(lowerCamelCase__ , lowerCamelCase__)) for vals in zip(scores.tolist() , lowerCamelCase__ , lowerCamelCase__) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel snake_case__ : List[str] = self.image_processor.post_process_object_detection(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__) snake_case__ : Optional[int] = raw_annotations[0] snake_case__ : Optional[Any] = raw_annotation["scores"] snake_case__ : Optional[int] = raw_annotation["labels"] snake_case__ : List[str] = raw_annotation["boxes"] snake_case__ : List[str] = scores.tolist() snake_case__ : Any = [self.model.config.idalabel[label.item()] for label in labels] snake_case__ : Optional[int] = [self._get_bounding_box(lowerCamelCase__) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] snake_case__ : Optional[int] = ["score", "label", "box"] snake_case__ : Optional[Any] = [ dict(zip(lowerCamelCase__ , lowerCamelCase__)) for vals in zip(raw_annotation["scores"] , raw_annotation["labels"] , raw_annotation["boxes"]) ] return annotation def UpperCAmelCase ( self , lowerCamelCase__) -> Dict[str, int]: '''simple docstring''' if self.framework != "pt": raise ValueError("The ObjectDetectionPipeline is only available in PyTorch.") snake_case__, snake_case__, snake_case__, snake_case__ : List[Any] = box.int().tolist() snake_case__ : Union[str, Any] = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
150
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowercase = logging.get_logger(__name__) def A__ ( _UpperCAmelCase : Optional[int] ) -> int: '''simple docstring''' snake_case__ : str = DPTConfig(embedding_type="hybrid" ) if "large" in checkpoint_url: snake_case__ : Union[str, Any] = 10_24 snake_case__ : Tuple = 40_96 snake_case__ : Any = 24 snake_case__ : Optional[Any] = 16 snake_case__ : Union[str, Any] = [5, 11, 17, 23] snake_case__ : List[str] = [2_56, 5_12, 10_24, 10_24] snake_case__ : int = (1, 3_84, 3_84) if "nyu" or "midas" in checkpoint_url: snake_case__ : Optional[Any] = 7_68 snake_case__ : Any = [1, 1, 1, 0.5] snake_case__ : str = [2_56, 5_12, 7_68, 7_68] snake_case__ : Tuple = 1_50 snake_case__ : Optional[Any] = 16 snake_case__ : Tuple = (1, 3_84, 3_84) snake_case__ : int = False snake_case__ : List[Any] = "project" if "ade" in checkpoint_url: snake_case__ : str = True snake_case__ : Optional[Any] = 7_68 snake_case__ : List[Any] = [1, 1, 1, 0.5] snake_case__ : Tuple = 1_50 snake_case__ : Tuple = 16 snake_case__ : Tuple = "huggingface/label-files" snake_case__ : Optional[Any] = "ade20k-id2label.json" snake_case__ : Any = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type="dataset" ) ) , "r" ) ) snake_case__ : Optional[Any] = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} snake_case__ : str = idalabel snake_case__ : Any = {v: k for k, v in idalabel.items()} snake_case__ : Any = [1, 1_50, 4_80, 4_80] return config, expected_shape def A__ ( _UpperCAmelCase : List[str] ) -> Union[str, Any]: '''simple docstring''' snake_case__ : str = ["pretrained.model.head.weight", "pretrained.model.head.bias"] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def A__ ( _UpperCAmelCase : Dict ) -> Any: '''simple docstring''' if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): snake_case__ : Optional[Any] = name.replace("pretrained.model" , "dpt.encoder" ) if "pretrained.model" in name: snake_case__ : Tuple = name.replace("pretrained.model" , "dpt.embeddings" ) if "patch_embed" in name: snake_case__ : int = name.replace("patch_embed" , "" ) if "pos_embed" in name: snake_case__ : Any = name.replace("pos_embed" , "position_embeddings" ) if "attn.proj" in name: snake_case__ : List[Any] = name.replace("attn.proj" , "attention.output.dense" ) if "proj" in name and "project" not in name: snake_case__ : List[Any] = name.replace("proj" , "projection" ) if "blocks" in name: snake_case__ : Dict = name.replace("blocks" , "layer" ) if "mlp.fc1" in name: snake_case__ : Any = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: snake_case__ : Optional[Any] = name.replace("mlp.fc2" , "output.dense" ) if "norm1" in name and "backbone" not in name: snake_case__ : Optional[int] = name.replace("norm1" , "layernorm_before" ) if "norm2" in name and "backbone" not in name: snake_case__ : List[str] = name.replace("norm2" , "layernorm_after" ) if "scratch.output_conv" in name: snake_case__ : Any = name.replace("scratch.output_conv" , "head" ) if "scratch" in name: snake_case__ : Union[str, Any] = name.replace("scratch" , "neck" ) if "layer1_rn" in name: snake_case__ : Optional[int] = name.replace("layer1_rn" , "convs.0" ) if "layer2_rn" in name: snake_case__ : str = name.replace("layer2_rn" , "convs.1" ) if "layer3_rn" in name: snake_case__ : Union[str, Any] = name.replace("layer3_rn" , "convs.2" ) if "layer4_rn" in name: snake_case__ : str = name.replace("layer4_rn" , "convs.3" ) if "refinenet" in name: snake_case__ : Union[str, Any] = int(name[len("neck.refinenet" ) : len("neck.refinenet" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 snake_case__ : List[str] = name.replace(F"""refinenet{layer_idx}""" , F"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: snake_case__ : Union[str, Any] = name.replace("out_conv" , "projection" ) if "resConfUnit1" in name: snake_case__ : Any = name.replace("resConfUnit1" , "residual_layer1" ) if "resConfUnit2" in name: snake_case__ : str = name.replace("resConfUnit2" , "residual_layer2" ) if "conv1" in name: snake_case__ : List[Any] = name.replace("conv1" , "convolution1" ) if "conv2" in name: snake_case__ : Tuple = name.replace("conv2" , "convolution2" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: snake_case__ : int = name.replace("pretrained.act_postprocess1.0.project.0" , "neck.reassemble_stage.readout_projects.0.0" ) if "pretrained.act_postprocess2.0.project.0" in name: snake_case__ : Dict = name.replace("pretrained.act_postprocess2.0.project.0" , "neck.reassemble_stage.readout_projects.1.0" ) if "pretrained.act_postprocess3.0.project.0" in name: snake_case__ : Tuple = name.replace("pretrained.act_postprocess3.0.project.0" , "neck.reassemble_stage.readout_projects.2.0" ) if "pretrained.act_postprocess4.0.project.0" in name: snake_case__ : Dict = name.replace("pretrained.act_postprocess4.0.project.0" , "neck.reassemble_stage.readout_projects.3.0" ) # resize blocks if "pretrained.act_postprocess1.3" in name: snake_case__ : Union[str, Any] = name.replace("pretrained.act_postprocess1.3" , "neck.reassemble_stage.layers.0.projection" ) if "pretrained.act_postprocess1.4" in name: snake_case__ : Dict = name.replace("pretrained.act_postprocess1.4" , "neck.reassemble_stage.layers.0.resize" ) if "pretrained.act_postprocess2.3" in name: snake_case__ : List[str] = name.replace("pretrained.act_postprocess2.3" , "neck.reassemble_stage.layers.1.projection" ) if "pretrained.act_postprocess2.4" in name: snake_case__ : str = name.replace("pretrained.act_postprocess2.4" , "neck.reassemble_stage.layers.1.resize" ) if "pretrained.act_postprocess3.3" in name: snake_case__ : Dict = name.replace("pretrained.act_postprocess3.3" , "neck.reassemble_stage.layers.2.projection" ) if "pretrained.act_postprocess4.3" in name: snake_case__ : Optional[Any] = name.replace("pretrained.act_postprocess4.3" , "neck.reassemble_stage.layers.3.projection" ) if "pretrained.act_postprocess4.4" in name: snake_case__ : Union[str, Any] = name.replace("pretrained.act_postprocess4.4" , "neck.reassemble_stage.layers.3.resize" ) if "pretrained" in name: snake_case__ : int = name.replace("pretrained" , "dpt" ) if "bn" in name: snake_case__ : Optional[Any] = name.replace("bn" , "batch_norm" ) if "head" in name: snake_case__ : int = name.replace("head" , "head.head" ) if "encoder.norm" in name: snake_case__ : List[Any] = name.replace("encoder.norm" , "layernorm" ) if "auxlayer" in name: snake_case__ : Optional[Any] = name.replace("auxlayer" , "auxiliary_head.head" ) if "backbone" in name: snake_case__ : Dict = name.replace("backbone" , "backbone.bit.encoder" ) if ".." in name: snake_case__ : Optional[int] = name.replace(".." , "." ) if "stem.conv" in name: snake_case__ : Optional[Any] = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: snake_case__ : List[str] = name.replace("blocks" , "layers" ) if "convolution" in name and "backbone" in name: snake_case__ : Optional[Any] = name.replace("convolution" , "conv" ) if "layer" in name and "backbone" in name: snake_case__ : Optional[int] = name.replace("layer" , "layers" ) if "backbone.bit.encoder.bit" in name: snake_case__ : str = name.replace("backbone.bit.encoder.bit" , "backbone.bit" ) if "embedder.conv" in name: snake_case__ : int = name.replace("embedder.conv" , "embedder.convolution" ) if "backbone.bit.encoder.stem.norm" in name: snake_case__ : Optional[int] = name.replace("backbone.bit.encoder.stem.norm" , "backbone.bit.embedder.norm" ) return name def A__ ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Any ) -> str: '''simple docstring''' for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case__ : str = state_dict.pop(F"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) snake_case__ : Optional[int] = state_dict.pop(F"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict snake_case__ : Dict = in_proj_weight[: config.hidden_size, :] snake_case__ : str = in_proj_bias[: config.hidden_size] snake_case__ : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case__ : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case__ : Optional[int] = in_proj_weight[ -config.hidden_size :, : ] snake_case__ : List[Any] = in_proj_bias[-config.hidden_size :] def A__ ( ) -> int: '''simple docstring''' snake_case__ : List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case__ : Dict = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def A__ ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : int , _UpperCAmelCase : List[str] , _UpperCAmelCase : Tuple , _UpperCAmelCase : List[str] ) -> Tuple: '''simple docstring''' snake_case__, snake_case__ : Tuple = get_dpt_config(_UpperCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") snake_case__ : Optional[int] = torch.load(_UpperCAmelCase , map_location="cpu" ) # remove certain keys remove_ignore_keys_(_UpperCAmelCase ) # rename keys for key in state_dict.copy().keys(): snake_case__ : Dict = state_dict.pop(_UpperCAmelCase ) snake_case__ : str = val # read in qkv matrices read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase ) # load HuggingFace model snake_case__ : Optional[int] = DPTForSemanticSegmentation(_UpperCAmelCase ) if "ade" in checkpoint_url else DPTForDepthEstimation(_UpperCAmelCase ) model.load_state_dict(_UpperCAmelCase ) model.eval() # Check outputs on an image snake_case__ : str = 4_80 if "ade" in checkpoint_url else 3_84 snake_case__ : int = DPTImageProcessor(size=_UpperCAmelCase ) snake_case__ : Tuple = prepare_img() snake_case__ : int = image_processor(_UpperCAmelCase , return_tensors="pt" ) # forward pass snake_case__ : Optional[Any] = model(**_UpperCAmelCase ).logits if "ade" in checkpoint_url else model(**_UpperCAmelCase ).predicted_depth if show_prediction: snake_case__ : Optional[int] = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode="bicubic" , align_corners=_UpperCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 2_55 ).show() if pytorch_dump_folder_path is not None: Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_UpperCAmelCase ) if push_to_hub: model.push_to_hub("ybelkada/dpt-hybrid-midas" ) image_processor.push_to_hub("ybelkada/dpt-hybrid-midas" ) if __name__ == "__main__": lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt""", type=str, help="""URL of the original DPT checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) parser.add_argument( """--model_name""", default="""dpt-large""", type=str, help="""Name of the model, in case you're pushing to the hub.""", ) parser.add_argument( """--show_prediction""", action="""store_true""", ) lowercase = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
150
1
'''simple docstring''' import baseaa def a__ ( _SCREAMING_SNAKE_CASE : str ) -> bytes: """simple docstring""" return baseaa.aaaencode(string.encode("utf-8" ) ) def a__ ( _SCREAMING_SNAKE_CASE : bytes ) -> str: """simple docstring""" return baseaa.aaadecode(_SCREAMING_SNAKE_CASE ).decode("utf-8" ) if __name__ == "__main__": import doctest doctest.testmod()
71
from __future__ import annotations _snake_case = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] _snake_case = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def lowerCAmelCase_ ( snake_case_ ): _A : str = [] _A : int = len(snake_case_ ) for i in range(snake_case_ ): _A : float = -1 for j in range(i + 1,snake_case_ ): if arr[i] < arr[j]: _A : Dict = arr[j] break result.append(snake_case_ ) return result def lowerCAmelCase_ ( snake_case_ ): _A : Optional[int] = [] for i, outer in enumerate(snake_case_ ): _A : float = -1 for inner in arr[i + 1 :]: if outer < inner: _A : List[str] = inner break result.append(snake_case_ ) return result def lowerCAmelCase_ ( snake_case_ ): _A : int = len(snake_case_ ) _A : list[float] = [] _A : list[float] = [-1] * arr_size for index in reversed(range(snake_case_ ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _A : Optional[int] = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) _snake_case = ( "from __main__ import arr, next_greatest_element_slow, " "next_greatest_element_fast, next_greatest_element" ) print( "next_greatest_element_slow():", timeit("next_greatest_element_slow(arr)", setup=setup), ) print( "next_greatest_element_fast():", timeit("next_greatest_element_fast(arr)", setup=setup), ) print( " next_greatest_element():", timeit("next_greatest_element(arr)", setup=setup), )
307
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A = logging.get_logger(__name__) def lowerCAmelCase__ ( lowerCamelCase__ ) -> Optional[int]: A = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: A = [144, 192, 240] A = [16, 32, 64, 96, 128, 160, 640] elif "mobilevit_xs" in mobilevit_name: A = [96, 120, 144] A = [16, 32, 48, 64, 80, 96, 384] elif "mobilevit_xxs" in mobilevit_name: A = [64, 80, 96] A = [16, 16, 24, 48, 64, 80, 320] A = 0.05 A = 2.0 if mobilevit_name.startswith('deeplabv3_' ): A = 512 A = 16 A = 21 A = 'pascal-voc-id2label.json' else: A = 1000 A = 'imagenet-1k-id2label.json' A = 'huggingface/label-files' A = json.load(open(hf_hub_download(lowerCamelCase__ , lowerCamelCase__ , repo_type='dataset' ) , 'r' ) ) A = {int(lowerCamelCase__ ): v for k, v in idalabel.items()} A = idalabel A = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase__ ( lowerCamelCase__ , lowerCamelCase__=False ) -> int: for i in range(1 , 6 ): if f"""layer_{i}.""" in name: A = name.replace(f"""layer_{i}.""" , f"""encoder.layer.{i - 1}.""" ) if "conv_1." in name: A = name.replace('conv_1.' , 'conv_stem.' ) if ".block." in name: A = name.replace('.block.' , '.' ) if "exp_1x1" in name: A = name.replace('exp_1x1' , 'expand_1x1' ) if "red_1x1" in name: A = name.replace('red_1x1' , 'reduce_1x1' ) if ".local_rep.conv_3x3." in name: A = name.replace('.local_rep.conv_3x3.' , '.conv_kxk.' ) if ".local_rep.conv_1x1." in name: A = name.replace('.local_rep.conv_1x1.' , '.conv_1x1.' ) if ".norm." in name: A = name.replace('.norm.' , '.normalization.' ) if ".conv." in name: A = name.replace('.conv.' , '.convolution.' ) if ".conv_proj." in name: A = name.replace('.conv_proj.' , '.conv_projection.' ) for i in range(0 , 2 ): for j in range(0 , 4 ): if f""".{i}.{j}.""" in name: A = name.replace(f""".{i}.{j}.""" , f""".{i}.layer.{j}.""" ) for i in range(2 , 6 ): for j in range(0 , 4 ): if f""".{i}.{j}.""" in name: A = name.replace(f""".{i}.{j}.""" , f""".{i}.""" ) if "expand_1x1" in name: A = name.replace('expand_1x1' , 'downsampling_layer.expand_1x1' ) if "conv_3x3" in name: A = name.replace('conv_3x3' , 'downsampling_layer.conv_3x3' ) if "reduce_1x1" in name: A = name.replace('reduce_1x1' , 'downsampling_layer.reduce_1x1' ) for i in range(2 , 5 ): if f""".global_rep.{i}.weight""" in name: A = name.replace(f""".global_rep.{i}.weight""" , '.layernorm.weight' ) if f""".global_rep.{i}.bias""" in name: A = name.replace(f""".global_rep.{i}.bias""" , '.layernorm.bias' ) if ".global_rep." in name: A = name.replace('.global_rep.' , '.transformer.' ) if ".pre_norm_mha.0." in name: A = name.replace('.pre_norm_mha.0.' , '.layernorm_before.' ) if ".pre_norm_mha.1.out_proj." in name: A = name.replace('.pre_norm_mha.1.out_proj.' , '.attention.output.dense.' ) if ".pre_norm_ffn.0." in name: A = name.replace('.pre_norm_ffn.0.' , '.layernorm_after.' ) if ".pre_norm_ffn.1." in name: A = name.replace('.pre_norm_ffn.1.' , '.intermediate.dense.' ) if ".pre_norm_ffn.4." in name: A = name.replace('.pre_norm_ffn.4.' , '.output.dense.' ) if ".transformer." in name: A = name.replace('.transformer.' , '.transformer.layer.' ) if ".aspp_layer." in name: A = name.replace('.aspp_layer.' , '.' ) if ".aspp_pool." in name: A = name.replace('.aspp_pool.' , '.' ) if "seg_head." in name: A = name.replace('seg_head.' , 'segmentation_head.' ) if "segmentation_head.classifier.classifier." in name: A = name.replace('segmentation_head.classifier.classifier.' , 'segmentation_head.classifier.' ) if "classifier.fc." in name: A = name.replace('classifier.fc.' , 'classifier.' ) elif (not base_model) and ("segmentation_head." not in name): A = 'mobilevit.' + name return name def lowerCAmelCase__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ) -> Any: if base_model: A = '' else: A = 'mobilevit.' for key in orig_state_dict.copy().keys(): A = orig_state_dict.pop(lowerCamelCase__ ) if key[:8] == "encoder.": A = key[8:] if "qkv" in key: A = key.split('.' ) A = int(key_split[0][6:] ) - 1 A = int(key_split[3] ) A = model.get_submodule(f"""{model_prefix}encoder.layer.{layer_num}""" ) A = layer.transformer.layer[transformer_num].attention.attention.all_head_size A = ( f"""{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.""" ) if "weight" in key: A = val[:dim, :] A = val[dim : dim * 2, :] A = val[-dim:, :] else: A = val[:dim] A = val[dim : dim * 2] A = val[-dim:] else: A = val return orig_state_dict def lowerCAmelCase__ ( ) -> Tuple: A = 'http://images.cocodataset.org/val2017/000000039769.jpg' A = Image.open(requests.get(lowerCamelCase__ , stream=lowerCamelCase__ ).raw ) return im @torch.no_grad() def lowerCAmelCase__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ) -> Union[str, Any]: A = get_mobilevit_config(lowerCamelCase__ ) # load original state_dict A = torch.load(lowerCamelCase__ , map_location='cpu' ) # load 🤗 model if mobilevit_name.startswith('deeplabv3_' ): A = MobileViTForSemanticSegmentation(lowerCamelCase__ ).eval() else: A = MobileViTForImageClassification(lowerCamelCase__ ).eval() A = convert_state_dict(lowerCamelCase__ , lowerCamelCase__ ) model.load_state_dict(lowerCamelCase__ ) # Check outputs on an image, prepared by MobileViTImageProcessor A = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) A = image_processor(images=prepare_img() , return_tensors='pt' ) A = model(**lowerCamelCase__ ) A = outputs.logits if mobilevit_name.startswith('deeplabv3_' ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": A = torch.tensor( [ [[6.20_65, 6.12_92, 6.20_70], [6.10_79, 6.12_54, 6.17_47], [6.00_42, 6.10_71, 6.10_34]], [[-6.92_53, -6.86_53, -7.03_98], [-7.32_18, -7.39_83, -7.36_70], [-7.19_61, -7.24_82, -7.15_69]], [[-4.47_23, -4.43_48, -4.37_69], [-5.36_29, -5.46_32, -5.45_98], [-5.15_87, -5.34_02, -5.50_59]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": A = torch.tensor( [ [[5.44_49, 5.57_33, 5.63_14], [5.18_15, 5.39_30, 5.59_63], [5.16_56, 5.43_33, 5.48_53]], [[-9.44_23, -9.77_66, -9.67_14], [-9.15_81, -9.57_20, -9.55_19], [-9.10_06, -9.64_58, -9.57_03]], [[-7.77_21, -7.37_16, -7.15_83], [-8.45_99, -8.06_24, -7.79_44], [-8.41_72, -7.83_66, -7.50_25]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": A = torch.tensor( [ [[6.98_11, 6.97_43, 7.31_23], [7.17_77, 7.19_31, 7.39_38], [7.56_33, 7.80_50, 7.89_01]], [[-10.55_36, -10.23_32, -10.29_24], [-10.23_36, -9.86_24, -9.59_64], [-10.88_40, -10.81_58, -10.66_59]], [[-3.49_38, -3.06_31, -2.86_20], [-3.42_05, -2.81_35, -2.68_75], [-3.41_79, -2.79_45, -2.87_50]], ] ) else: raise ValueError(f"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase__ , atol=1E-4 ) else: assert logits.shape == (1, 1000) if mobilevit_name == "mobilevit_s": A = torch.tensor([-0.98_66, 0.23_92, -1.12_41] ) elif mobilevit_name == "mobilevit_xs": A = torch.tensor([-2.47_61, -0.93_99, -1.95_87] ) elif mobilevit_name == "mobilevit_xxs": A = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ) else: raise ValueError(f"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3] , lowerCamelCase__ , atol=1E-4 ) Path(lowerCamelCase__ ).mkdir(exist_ok=lowerCamelCase__ ) print(f"""Saving model {mobilevit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowerCamelCase__ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(lowerCamelCase__ ) if push_to_hub: A = { 'mobilevit_s': 'mobilevit-small', 'mobilevit_xs': 'mobilevit-x-small', 'mobilevit_xxs': 'mobilevit-xx-small', 'deeplabv3_mobilevit_s': 'deeplabv3-mobilevit-small', 'deeplabv3_mobilevit_xs': 'deeplabv3-mobilevit-x-small', 'deeplabv3_mobilevit_xxs': 'deeplabv3-mobilevit-xx-small', } print('Pushing to the hub...' ) A = model_mapping[mobilevit_name] image_processor.push_to_hub(lowerCamelCase__ , organization='apple' ) model.push_to_hub(lowerCamelCase__ , organization='apple' ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--mobilevit_name', default='mobilevit_s', type=str, help=( 'Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\',' ' \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'.' ), ) parser.add_argument( '--checkpoint_path', required=True, type=str, help='Path to the original state dict (.pt file).' ) parser.add_argument( '--pytorch_dump_folder_path', required=True, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) A = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
109
"""simple docstring""" import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging A = logging.get_logger(__name__) def lowerCAmelCase__ ( lowerCamelCase__=None , lowerCamelCase__=None ) -> Dict: return field(default_factory=lambda: default , metadata=lowerCamelCase__ ) @dataclass class UpperCAmelCase__ : lowerCAmelCase_ : List[str] = list_field( default=[] ,metadata={ """help""": ( """Model checkpoints to be provided to the AutoModel classes. Leave blank to benchmark the base version""" """ of all available models""" ) } ,) lowerCAmelCase_ : List[int] = list_field( default=[8] ,metadata={"""help""": """List of batch sizes for which memory and time performance will be evaluated"""} ) lowerCAmelCase_ : List[int] = list_field( default=[8, 32, 1_28, 5_12] ,metadata={"""help""": """List of sequence lengths for which memory and time performance will be evaluated"""} ,) lowerCAmelCase_ : bool = field( default=UpperCamelCase ,metadata={"""help""": """Whether to benchmark inference of model. Inference can be disabled via --no-inference."""} ,) lowerCAmelCase_ : bool = field( default=UpperCamelCase ,metadata={"""help""": """Whether to run on available cuda devices. Cuda can be disabled via --no-cuda."""} ,) lowerCAmelCase_ : bool = field( default=UpperCamelCase ,metadata={"""help""": """Whether to run on available tpu devices. TPU can be disabled via --no-tpu."""} ) lowerCAmelCase_ : bool = field(default=UpperCamelCase ,metadata={"""help""": """Use FP16 to accelerate inference."""} ) lowerCAmelCase_ : bool = field(default=UpperCamelCase ,metadata={"""help""": """Benchmark training of model"""} ) lowerCAmelCase_ : bool = field(default=UpperCamelCase ,metadata={"""help""": """Verbose memory tracing"""} ) lowerCAmelCase_ : bool = field( default=UpperCamelCase ,metadata={"""help""": """Whether to perform speed measurements. Speed measurements can be disabled via --no-speed."""} ,) lowerCAmelCase_ : bool = field( default=UpperCamelCase ,metadata={ """help""": """Whether to perform memory measurements. Memory measurements can be disabled via --no-memory""" } ,) lowerCAmelCase_ : bool = field(default=UpperCamelCase ,metadata={"""help""": """Trace memory line by line"""} ) lowerCAmelCase_ : bool = field(default=UpperCamelCase ,metadata={"""help""": """Save result to a CSV file"""} ) lowerCAmelCase_ : bool = field(default=UpperCamelCase ,metadata={"""help""": """Save all print statements in a log file"""} ) lowerCAmelCase_ : bool = field(default=UpperCamelCase ,metadata={"""help""": """Whether to print environment information"""} ) lowerCAmelCase_ : bool = field( default=UpperCamelCase ,metadata={ """help""": ( """Whether to use multiprocessing for memory and speed measurement. It is highly recommended to use""" """ multiprocessing for accurate CPU and GPU memory measurements. This option should only be disabled""" """ for debugging / testing and on TPU.""" ) } ,) lowerCAmelCase_ : str = field( default=f"""inference_time_{round(time() )}.csv""" ,metadata={"""help""": """CSV filename used if saving time results to csv."""} ,) lowerCAmelCase_ : str = field( default=f"""inference_memory_{round(time() )}.csv""" ,metadata={"""help""": """CSV filename used if saving memory results to csv."""} ,) lowerCAmelCase_ : str = field( default=f"""train_time_{round(time() )}.csv""" ,metadata={"""help""": """CSV filename used if saving time results to csv for training."""} ,) lowerCAmelCase_ : str = field( default=f"""train_memory_{round(time() )}.csv""" ,metadata={"""help""": """CSV filename used if saving memory results to csv for training."""} ,) lowerCAmelCase_ : str = field( default=f"""env_info_{round(time() )}.csv""" ,metadata={"""help""": """CSV filename used if saving environment information."""} ,) lowerCAmelCase_ : str = field( default=f"""log_{round(time() )}.csv""" ,metadata={"""help""": """Log filename used if print statements are saved in log."""} ,) lowerCAmelCase_ : int = field(default=3 ,metadata={"""help""": """Times an experiment will be run."""} ) lowerCAmelCase_ : bool = field( default=UpperCamelCase ,metadata={ """help""": ( """Instead of loading the model as defined in `config.architectures` if exists, just load the pretrain""" """ model weights.""" ) } ,) def A_ ( self : Dict ) -> Optional[int]: '''simple docstring''' warnings.warn( f"""The class {self.__class__} is deprecated. Hugging Face Benchmarking utils""" ' are deprecated in general and it is advised to use external Benchmarking libraries ' ' to benchmark Transformer models.' , snake_case , ) def A_ ( self : str ) -> List[Any]: '''simple docstring''' return json.dumps(dataclasses.asdict(self ) , indent=2 ) @property def A_ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' if len(self.models ) <= 0: raise ValueError( 'Please make sure you provide at least one model name / model identifier, *e.g.* `--models' ' bert-base-cased` or `args.models = [\'bert-base-cased\'].' ) return self.models @property def A_ ( self : Tuple ) -> List[str]: '''simple docstring''' if not self.multi_process: return False elif self.is_tpu: logger.info('Multiprocessing is currently not possible on TPU.' ) return False else: return True
109
1
"""simple docstring""" from __future__ import annotations def _UpperCamelCase ( _A , _A , _A , ) -> tuple: """simple docstring""" if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif electron_conc < 0: raise ValueError("""Electron concentration cannot be negative in a semiconductor""" ) elif hole_conc < 0: raise ValueError("""Hole concentration cannot be negative in a semiconductor""" ) elif intrinsic_conc < 0: raise ValueError( """Intrinsic concentration cannot be negative in a semiconductor""" ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
555
"""simple docstring""" import cva import numpy as np class a_ : def __init__( self : Optional[Any] , __UpperCamelCase : float , __UpperCamelCase : int ) ->Dict: '''simple docstring''' if k in (0.0_4, 0.0_6): _UpperCAmelCase = k _UpperCAmelCase = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : Tuple ) ->str: '''simple docstring''' return str(self.k ) def _snake_case ( self : str , __UpperCamelCase : str ) ->tuple[cva.Mat, list[list[int]]]: '''simple docstring''' _UpperCAmelCase = cva.imread(__UpperCamelCase , 0 ) _UpperCAmelCase ,_UpperCAmelCase = img.shape _UpperCAmelCase = [] _UpperCAmelCase = img.copy() _UpperCAmelCase = cva.cvtColor(__UpperCamelCase , cva.COLOR_GRAY2RGB ) _UpperCAmelCase ,_UpperCAmelCase = np.gradient(__UpperCamelCase ) _UpperCAmelCase = dx**2 _UpperCAmelCase = dy**2 _UpperCAmelCase = dx * dy _UpperCAmelCase = 0.0_4 _UpperCAmelCase = self.window_size // 2 for y in range(__UpperCamelCase , h - offset ): for x in range(__UpperCamelCase , w - offset ): _UpperCAmelCase = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _UpperCAmelCase = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _UpperCAmelCase = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _UpperCAmelCase = (wxx * wyy) - (wxy**2) _UpperCAmelCase = wxx + wyy _UpperCAmelCase = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 2_55 ) return color_img, corner_list if __name__ == "__main__": a : List[Any] = HarrisCorner(0.04, 3) a , a : List[Any] = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
555
1
'''simple docstring''' _lowerCamelCase = "0.21.0" from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
710
'''simple docstring''' from __future__ import annotations from typing import Any def a__ ( _SCREAMING_SNAKE_CASE : list ) -> int: """simple docstring""" if not postfix_notation: return 0 UpperCAmelCase_ : Tuple = {"+", "-", "*", "/"} UpperCAmelCase_ : list[Any] = [] for token in postfix_notation: if token in operations: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = stack.pop(), stack.pop() if token == "+": stack.append(a + b ) elif token == "-": stack.append(a - b ) elif token == "*": stack.append(a * b ) else: if a * b < 0 and a % b != 0: stack.append(a // b + 1 ) else: stack.append(a // b ) else: stack.append(int(_SCREAMING_SNAKE_CASE ) ) return stack.pop() if __name__ == "__main__": import doctest doctest.testmod()
323
0
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __A = "\\n@inproceedings{popovic-2015-chrf,\n title = \"chr{F}: character n-gram {F}-score for automatic {MT} evaluation\",\n author = \"Popovi{\'c}, Maja\",\n booktitle = \"Proceedings of the Tenth Workshop on Statistical Machine Translation\",\n month = sep,\n year = \"2015\",\n address = \"Lisbon, Portugal\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/W15-3049\",\n doi = \"10.18653/v1/W15-3049\",\n pages = \"392--395\",\n}\n@inproceedings{popovic-2017-chrf,\n title = \"chr{F}++: words helping character n-grams\",\n author = \"Popovi{\'c}, Maja\",\n booktitle = \"Proceedings of the Second Conference on Machine Translation\",\n month = sep,\n year = \"2017\",\n address = \"Copenhagen, Denmark\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/W17-4770\",\n doi = \"10.18653/v1/W17-4770\",\n pages = \"612--618\",\n}\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" __A = "\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n" __A = "\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n 'score' (float): The chrF (chrF++) score,\n 'char_order' (int): The character n-gram order,\n 'word_order' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n 'beta' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {'score': 84.64214891738334, 'char_order': 6, 'word_order': 0, 'beta': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {'score': 82.87263732906315, 'char_order': 6, 'word_order': 2, 'beta': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {'score': 92.12853119829202, 'char_order': 6, 'word_order': 2, 'beta': 2}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A ( datasets.Metric ): def A__ ( self ) -> Union[str, Any]: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/mjpost/sacreBLEU#chrf--chrf""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#chrf--chrf"""] , reference_urls=[ """https://github.com/m-popovic/chrF""", ] , ) def A__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = CHRF.CHAR_ORDER , lowerCamelCase__ = CHRF.WORD_ORDER , lowerCamelCase__ = CHRF.BETA , lowerCamelCase__ = False , lowerCamelCase__ = False , lowerCamelCase__ = False , ) -> Any: '''simple docstring''' lowercase__ = len(references[0] ) if any(len(lowerCamelCase__ ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) lowercase__ = [[refs[i] for refs in references] for i in range(lowerCamelCase__ )] lowercase__ = CHRF(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) lowercase__ = sb_chrf.corpus_score(lowerCamelCase__ , lowerCamelCase__ ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
325
'''simple docstring''' import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") __A = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) __A = requests.get(url, headers={"UserAgent": UserAgent().random}) # res.raise_for_status() with open("project1a.html", "wb") as out_file: # only for knowing the class for data in res.iter_content(10_000): out_file.write(data) __A = BeautifulSoup(res.text, "html.parser") __A = list(soup.select(".eZt8xd"))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get("href")) else: webbrowser.open(F'''https://google.com{link.get("href")}''')
325
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer a__ = logging.get_logger(__name__) a__ = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} a__ = { """vocab_file""": {"""mobilebert-uncased""": """https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt"""}, """tokenizer_file""": { """mobilebert-uncased""": """https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json""" }, } a__ = {"""mobilebert-uncased""": 5_1_2} a__ = {} class _lowerCAmelCase ( lowercase_ ): """simple docstring""" _lowercase : List[str] = VOCAB_FILES_NAMES _lowercase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP _lowercase : Union[str, Any] = PRETRAINED_INIT_CONFIGURATION _lowercase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase : str = MobileBertTokenizer def __init__( self : Dict , UpperCamelCase__ : int=None , UpperCamelCase__ : str=None , UpperCamelCase__ : int=True , UpperCamelCase__ : str="[UNK]" , UpperCamelCase__ : Tuple="[SEP]" , UpperCamelCase__ : Optional[Any]="[PAD]" , UpperCamelCase__ : int="[CLS]" , UpperCamelCase__ : List[str]="[MASK]" , UpperCamelCase__ : Dict=True , UpperCamelCase__ : Optional[int]=None , **UpperCamelCase__ : Optional[Any] , ): '''simple docstring''' super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , do_lower_case=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , tokenize_chinese_chars=UpperCamelCase__ , strip_accents=UpperCamelCase__ , **UpperCamelCase__ , ) snake_case__ = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get("""lowercase""" , UpperCamelCase__) != do_lower_case or normalizer_state.get("""strip_accents""" , UpperCamelCase__) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , UpperCamelCase__) != tokenize_chinese_chars ): snake_case__ = getattr(UpperCamelCase__ , normalizer_state.pop("""type""")) snake_case__ = do_lower_case snake_case__ = strip_accents snake_case__ = tokenize_chinese_chars snake_case__ = normalizer_class(**UpperCamelCase__) snake_case__ = do_lower_case def __magic_name__ ( self : Dict , UpperCamelCase__ : str , UpperCamelCase__ : Optional[Any]=None): '''simple docstring''' snake_case__ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __magic_name__ ( self : Tuple , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None): '''simple docstring''' snake_case__ = [self.sep_token_id] snake_case__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def __magic_name__ ( self : Optional[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None): '''simple docstring''' snake_case__ = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__) return tuple(UpperCamelCase__)
99
import math from collections.abc import Iterator from itertools import takewhile def _UpperCAmelCase ( a : int ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(a ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _UpperCAmelCase ( ): snake_case__ = 2 while True: if is_prime(a ): yield num num += 1 def _UpperCAmelCase ( a : int = 200_0000 ): return sum(takewhile(lambda a : x < n , prime_generator() ) ) if __name__ == "__main__": print(F'''{solution() = }''')
99
1
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 lowerCAmelCase :Optional[int] = get_tests_dir('''fixtures''') lowerCAmelCase :Any = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') lowerCAmelCase :Tuple = get_tests_dir('''fixtures/dummy-config.json''') class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : Union[str, Any] ) -> int: __magic_name__ : str = 0 def __lowerCAmelCase ( self : int ) -> Tuple: __magic_name__ : List[str] = AutoFeatureExtractor.from_pretrained('facebook/wav2vec2-base-960h' ) self.assertIsInstance(_A , _A ) def __lowerCAmelCase ( self : Tuple ) -> List[Any]: __magic_name__ : List[str] = AutoFeatureExtractor.from_pretrained(_A ) self.assertIsInstance(_A , _A ) def __lowerCAmelCase ( self : Tuple ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ : Union[str, Any] = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __magic_name__ : Dict = AutoFeatureExtractor.from_pretrained(_A ).to_dict() config_dict.pop('feature_extractor_type' ) __magic_name__ : int = WavaVecaFeatureExtractor(**_A ) # save in new folder model_config.save_pretrained(_A ) config.save_pretrained(_A ) __magic_name__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained(_A ) # make sure private variable is not incorrectly saved __magic_name__ : List[str] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(_A , _A ) def __lowerCAmelCase ( self : int ) -> Union[str, Any]: __magic_name__ : Tuple = AutoFeatureExtractor.from_pretrained(_A ) self.assertIsInstance(_A , _A ) def __lowerCAmelCase ( self : List[str] ) -> Optional[int]: with self.assertRaisesRegex( _A , 'bert-base is not a local folder and is not a valid model identifier' ): __magic_name__ : str = AutoFeatureExtractor.from_pretrained('bert-base' ) def __lowerCAmelCase ( self : Any ) -> Tuple: with self.assertRaisesRegex( _A , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): __magic_name__ : Tuple = AutoFeatureExtractor.from_pretrained(_A , revision='aaaaaa' ) def __lowerCAmelCase ( self : Dict ) -> str: with self.assertRaisesRegex( _A , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): __magic_name__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained('hf-internal-testing/config-no-model' ) def __lowerCAmelCase ( self : Optional[int] ) -> Union[str, Any]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_A ): __magic_name__ : Dict = AutoFeatureExtractor.from_pretrained( 'hf-internal-testing/test_dynamic_feature_extractor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(_A ): __magic_name__ : Optional[Any] = AutoFeatureExtractor.from_pretrained( 'hf-internal-testing/test_dynamic_feature_extractor' , trust_remote_code=_A ) __magic_name__ : List[Any] = AutoFeatureExtractor.from_pretrained( 'hf-internal-testing/test_dynamic_feature_extractor' , trust_remote_code=_A ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(_A ) __magic_name__ : List[Any] = AutoFeatureExtractor.from_pretrained(_A , trust_remote_code=_A ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) def __lowerCAmelCase ( self : str ) -> Tuple: try: AutoConfig.register('custom' , _A ) AutoFeatureExtractor.register(_A , _A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_A ): AutoFeatureExtractor.register(_A , _A ) # Now that the config is registered, it can be used as any other config with the auto-API __magic_name__ : str = CustomFeatureExtractor.from_pretrained(_A ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(_A ) __magic_name__ : List[Any] = AutoFeatureExtractor.from_pretrained(_A ) self.assertIsInstance(_A , _A ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def __lowerCAmelCase ( self : Optional[Any] ) -> Dict: class _lowerCamelCase ( lowercase__ ): '''simple docstring''' A_ : Any = True try: AutoConfig.register('custom' , _A ) AutoFeatureExtractor.register(_A , _A ) # If remote code is not set, the default is to use local __magic_name__ : Optional[Any] = AutoFeatureExtractor.from_pretrained( 'hf-internal-testing/test_dynamic_feature_extractor' ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __magic_name__ : str = AutoFeatureExtractor.from_pretrained( 'hf-internal-testing/test_dynamic_feature_extractor' , trust_remote_code=_A ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __magic_name__ : Tuple = AutoFeatureExtractor.from_pretrained( 'hf-internal-testing/test_dynamic_feature_extractor' , trust_remote_code=_A ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) self.assertTrue(not hasattr(_A , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
561
'''simple docstring''' from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder lowerCAmelCase :Dict = datasets.utils.logging.get_logger(__name__) class _lowerCamelCase ( folder_based_builder.FolderBasedBuilderConfig ): '''simple docstring''' A_ : bool = None A_ : bool = None class _lowerCamelCase ( folder_based_builder.FolderBasedBuilder ): '''simple docstring''' A_ : Union[str, Any] = datasets.Audio() A_ : Tuple = """audio""" A_ : Optional[Any] = AudioFolderConfig A_ : List[str] # definition at the bottom of the script A_ : Any = AudioClassification(audio_column="""audio""" , label_column="""label""" ) lowerCAmelCase :List[str] = [ '''.aiff''', '''.au''', '''.avr''', '''.caf''', '''.flac''', '''.htk''', '''.svx''', '''.mat4''', '''.mat5''', '''.mpc2k''', '''.ogg''', '''.paf''', '''.pvf''', '''.raw''', '''.rf64''', '''.sd2''', '''.sds''', '''.ircam''', '''.voc''', '''.w64''', '''.wav''', '''.nist''', '''.wavex''', '''.wve''', '''.xi''', '''.mp3''', '''.opus''', ] lowerCAmelCase :str = AUDIO_EXTENSIONS
561
1
"""simple docstring""" from __future__ import annotations def lowercase_ ( _lowercase : list[int] ): # This function is recursive '''simple docstring''' UpperCAmelCase : Optional[int] = len(snake_case__ ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else UpperCAmelCase : Dict = array[0] UpperCAmelCase : List[Any] = False UpperCAmelCase : Any = 1 UpperCAmelCase : list[int] = [] while not is_found and i < array_length: if array[i] < pivot: UpperCAmelCase : int = True UpperCAmelCase : str = [element for element in array[i:] if element >= array[i]] UpperCAmelCase : Optional[int] = longest_subsequence(snake_case__ ) if len(snake_case__ ) > len(snake_case__ ): UpperCAmelCase : str = temp_array else: i += 1 UpperCAmelCase : Optional[int] = [element for element in array[1:] if element >= pivot] UpperCAmelCase : Union[str, Any] = [pivot, *longest_subsequence(snake_case__ )] if len(snake_case__ ) > len(snake_case__ ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
706
"""simple docstring""" from typing import Any class snake_case__ : def __init__( self : Union[str, Any] , lowercase : Any ): '''simple docstring''' UpperCAmelCase : Dict = data UpperCAmelCase : Optional[Any] = None def __repr__( self : str ): '''simple docstring''' return f"""Node({self.data})""" class snake_case__ : def __init__( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase : Optional[Any] = None def __iter__( self : Dict ): '''simple docstring''' UpperCAmelCase : str = self.head while node: yield node.data UpperCAmelCase : Tuple = node.next def __len__( self : Tuple ): '''simple docstring''' return sum(1 for _ in self ) def __repr__( self : Any ): '''simple docstring''' return "->".join([str(lowercase ) for item in self] ) def __getitem__( self : List[str] , lowercase : int ): '''simple docstring''' if not 0 <= index < len(self ): raise ValueError("list index out of range." ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : str , lowercase : int , lowercase : Any ): '''simple docstring''' if not 0 <= index < len(self ): raise ValueError("list index out of range." ) UpperCAmelCase : List[str] = self.head for _ in range(lowercase ): UpperCAmelCase : List[Any] = current.next UpperCAmelCase : List[str] = data def __lowerCAmelCase ( self : Optional[Any] , lowercase : Any ): '''simple docstring''' self.insert_nth(len(self ) , lowercase ) def __lowerCAmelCase ( self : List[str] , lowercase : Any ): '''simple docstring''' self.insert_nth(0 , lowercase ) def __lowerCAmelCase ( self : List[str] , lowercase : int , lowercase : Any ): '''simple docstring''' if not 0 <= index <= len(self ): raise IndexError("list index out of range" ) UpperCAmelCase : List[Any] = Node(lowercase ) if self.head is None: UpperCAmelCase : Optional[Any] = new_node elif index == 0: UpperCAmelCase : Optional[int] = self.head # link new_node to head UpperCAmelCase : Optional[Any] = new_node else: UpperCAmelCase : Optional[int] = self.head for _ in range(index - 1 ): UpperCAmelCase : Union[str, Any] = temp.next UpperCAmelCase : Dict = temp.next UpperCAmelCase : Any = new_node def __lowerCAmelCase ( self : Dict ): # print every node data '''simple docstring''' print(self ) def __lowerCAmelCase ( self : Optional[Any] ): '''simple docstring''' return self.delete_nth(0 ) def __lowerCAmelCase ( self : str ): # delete from tail '''simple docstring''' return self.delete_nth(len(self ) - 1 ) def __lowerCAmelCase ( self : Optional[Any] , lowercase : int = 0 ): '''simple docstring''' if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("List index out of range." ) UpperCAmelCase : List[str] = self.head # default first node if index == 0: UpperCAmelCase : Tuple = self.head.next else: UpperCAmelCase : Union[str, Any] = self.head for _ in range(index - 1 ): UpperCAmelCase : Dict = temp.next UpperCAmelCase : Optional[Any] = temp.next UpperCAmelCase : int = temp.next.next return delete_node.data def __lowerCAmelCase ( self : List[Any] ): '''simple docstring''' return self.head is None def __lowerCAmelCase ( self : int ): '''simple docstring''' UpperCAmelCase : List[Any] = None UpperCAmelCase : str = self.head while current: # Store the current node's next node. UpperCAmelCase : List[Any] = current.next # Make the current node's next point backwards UpperCAmelCase : List[str] = prev # Make the previous node be the current node UpperCAmelCase : Optional[Any] = current # Make the current node the next node (to progress iteration) UpperCAmelCase : Any = next_node # Return prev in order to put the head at the end UpperCAmelCase : List[str] = prev def lowercase_ ( ): '''simple docstring''' UpperCAmelCase : Optional[int] = LinkedList() assert linked_list.is_empty() is True assert str(_lowercase ) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10 ): assert len(_lowercase ) == i linked_list.insert_nth(_lowercase , i + 1 ) assert str(_lowercase ) == "->".join(str(_lowercase ) for i in range(1 , 11 ) ) linked_list.insert_head(0 ) linked_list.insert_tail(11 ) assert str(_lowercase ) == "->".join(str(_lowercase ) for i in range(0 , 12 ) ) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9 ) == 10 assert linked_list.delete_tail() == 11 assert len(_lowercase ) == 9 assert str(_lowercase ) == "->".join(str(_lowercase ) for i in range(1 , 10 ) ) assert all(linked_list[i] == i + 1 for i in range(0 , 9 ) ) is True for i in range(0 , 9 ): UpperCAmelCase : Union[str, Any] = -i assert all(linked_list[i] == -i for i in range(0 , 9 ) ) is True linked_list.reverse() assert str(_lowercase ) == "->".join(str(_lowercase ) for i in range(-8 , 1 ) ) def lowercase_ ( ): '''simple docstring''' UpperCAmelCase : Optional[Any] = [ -9, 1_00, Node(77_34_51_12 ), "dlrow olleH", 7, 55_55, 0, -1_9_2.5_5_5_5_5, "Hello, world!", 7_7.9, Node(10 ), None, None, 1_2.2_0, ] UpperCAmelCase : Optional[Any] = LinkedList() for i in test_input: linked_list.insert_tail(_lowercase ) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(_lowercase ) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head UpperCAmelCase : Tuple = linked_list.delete_head() assert result == -9 assert ( str(_lowercase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail UpperCAmelCase : Any = linked_list.delete_tail() assert result == 1_2.2 assert ( str(_lowercase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list UpperCAmelCase : Union[str, Any] = linked_list.delete_nth(10 ) assert result is None assert ( str(_lowercase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("Hello again, world!" ) ) assert ( str(_lowercase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(_lowercase ) assert ( str(_lowercase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(_lowercase ) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def lowercase_ ( ): '''simple docstring''' from doctest import testmod testmod() UpperCAmelCase : Any = LinkedList() linked_list.insert_head(input("Inserting 1st at head " ).strip() ) linked_list.insert_head(input("Inserting 2nd at head " ).strip() ) print("\nPrint list:" ) linked_list.print_list() linked_list.insert_tail(input("\nInserting 1st at tail " ).strip() ) linked_list.insert_tail(input("Inserting 2nd at tail " ).strip() ) print("\nPrint list:" ) linked_list.print_list() print("\nDelete head" ) linked_list.delete_head() print("Delete tail" ) linked_list.delete_tail() print("\nPrint list:" ) linked_list.print_list() print("\nReverse linked list" ) linked_list.reverse() print("\nPrint list:" ) linked_list.print_list() print("\nString representation of linked list:" ) print(_lowercase ) print("\nReading/changing Node data using indexing:" ) print(F"""Element at Position 1: {linked_list[1]}""" ) UpperCAmelCase : int = input("Enter New Value: " ).strip() print("New list:" ) print(_lowercase ) print(F"""length of linked_list is : {len(_lowercase )}""" ) if __name__ == "__main__": main()
292
0
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union _a : str = TypeVar("T") _a : Dict = Union[List[T], Tuple[T, ...]] _a : str = Union[T, List[T], Dict[str, T]] _a : Union[str, Any] = Union[str, bytes, os.PathLike]
56
'''simple docstring''' def _a (lowercase__ : int , lowercase__ : int , lowercase__ : list[list[int]] ) -> int: """simple docstring""" def update_area_of_max_square(lowercase__ : int , lowercase__ : int ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __snake_case = update_area_of_max_square(lowercase__ , col + 1 ) __snake_case = update_area_of_max_square(row + 1 , col + 1 ) __snake_case = update_area_of_max_square(row + 1 , lowercase__ ) if mat[row][col]: __snake_case = 1 + min([right, diagonal, down] ) __snake_case = max(largest_square_area[0] , lowercase__ ) return sub_problem_sol else: return 0 __snake_case = [0] update_area_of_max_square(0 , 0 ) return largest_square_area[0] def _a (lowercase__ : int , lowercase__ : int , lowercase__ : list[list[int]] ) -> int: """simple docstring""" def update_area_of_max_square_using_dp_array( lowercase__ : int , lowercase__ : int , lowercase__ : list[list[int]] ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __snake_case = update_area_of_max_square_using_dp_array(lowercase__ , col + 1 , lowercase__ ) __snake_case = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , lowercase__ ) __snake_case = update_area_of_max_square_using_dp_array(row + 1 , lowercase__ , lowercase__ ) if mat[row][col]: __snake_case = 1 + min([right, diagonal, down] ) __snake_case = max(largest_square_area[0] , lowercase__ ) __snake_case = sub_problem_sol return sub_problem_sol else: return 0 __snake_case = [0] __snake_case = [[-1] * cols for _ in range(lowercase__ )] update_area_of_max_square_using_dp_array(0 , 0 , lowercase__ ) return largest_square_area[0] def _a (lowercase__ : int , lowercase__ : int , lowercase__ : list[list[int]] ) -> int: """simple docstring""" __snake_case = [[0] * (cols + 1) for _ in range(rows + 1 )] __snake_case = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __snake_case = dp_array[row][col + 1] __snake_case = dp_array[row + 1][col + 1] __snake_case = dp_array[row + 1][col] if mat[row][col] == 1: __snake_case = 1 + min(lowercase__ , lowercase__ , lowercase__ ) __snake_case = max(dp_array[row][col] , lowercase__ ) else: __snake_case = 0 return largest_square_area def _a (lowercase__ : int , lowercase__ : int , lowercase__ : list[list[int]] ) -> int: """simple docstring""" __snake_case = [0] * (cols + 1) __snake_case = [0] * (cols + 1) __snake_case = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __snake_case = current_row[col + 1] __snake_case = next_row[col + 1] __snake_case = next_row[col] if mat[row][col] == 1: __snake_case = 1 + min(lowercase__ , lowercase__ , lowercase__ ) __snake_case = max(current_row[col] , lowercase__ ) else: __snake_case = 0 __snake_case = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
56
1
import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : List[str] = os.path.abspath(snake_case_ ) logger.info(f'''Converting TensorFlow checkpoint from {tf_path}''' ) # Load weights from TF model _A : List[str] = tf.train.list_variables(snake_case_ ) _A : List[Any] = [] _A : Optional[Any] = [] _A : Union[str, Any] = [] for full_name, shape in init_vars: # logger.info(f"Loading TF weight {name} with shape {shape}") _A : Tuple = full_name.split("""/""" ) if full_name == "_CHECKPOINTABLE_OBJECT_GRAPH" or name[0] in ["global_step", "save_counter"]: logger.info(f'''Skipping non-model layer {full_name}''' ) continue if "optimizer" in full_name: logger.info(f'''Skipping optimization layer {full_name}''' ) continue if name[0] == "model": # ignore initial 'model' _A : int = name[1:] # figure out how many levels deep the name is _A : Any = 0 for _name in name: if _name.startswith("""layer_with_weights""" ): depth += 1 else: break layer_depth.append(snake_case_ ) # read data _A : List[str] = tf.train.load_variable(snake_case_,snake_case_ ) names.append("""/""".join(snake_case_ ) ) arrays.append(snake_case_ ) logger.info(f'''Read a total of {len(snake_case_ ):,} layers''' ) # Sanity check if len(set(snake_case_ ) ) != 1: raise ValueError(f'''Found layer names with different depths (layer depth {list(set(snake_case_ ) )})''' ) _A : List[str] = list(set(snake_case_ ) )[0] if layer_depth != 1: raise ValueError( """The model contains more than just the embedding/encoder layers. This script does not handle MLM/NSP""" """ heads.""" ) # convert layers logger.info("""Converting weights...""" ) for full_name, array in zip(snake_case_,snake_case_ ): _A : Optional[Any] = full_name.split("""/""" ) _A : Any = model _A : Any = [] for i, m_name in enumerate(snake_case_ ): if m_name == ".ATTRIBUTES": # variable names end with .ATTRIBUTES/VARIABLE_VALUE break if m_name.startswith("""layer_with_weights""" ): _A : Optional[Any] = int(m_name.split("""-""" )[-1] ) if layer_num <= 2: # embedding layers # layer_num 0: word_embeddings # layer_num 1: position_embeddings # layer_num 2: token_type_embeddings continue elif layer_num == 3: # embedding LayerNorm trace.extend(["""embeddings""", """LayerNorm"""] ) _A : int = getattr(snake_case_,"""embeddings""" ) _A : Dict = getattr(snake_case_,"""LayerNorm""" ) elif layer_num > 3 and layer_num < config.num_hidden_layers + 4: # encoder layers trace.extend(["""encoder""", """layer""", str(layer_num - 4 )] ) _A : Any = getattr(snake_case_,"""encoder""" ) _A : Tuple = getattr(snake_case_,"""layer""" ) _A : str = pointer[layer_num - 4] elif layer_num == config.num_hidden_layers + 4: # pooler layer trace.extend(["""pooler""", """dense"""] ) _A : str = getattr(snake_case_,"""pooler""" ) _A : Union[str, Any] = getattr(snake_case_,"""dense""" ) elif m_name == "embeddings": trace.append("""embeddings""" ) _A : List[str] = getattr(snake_case_,"""embeddings""" ) if layer_num == 0: trace.append("""word_embeddings""" ) _A : Optional[Any] = getattr(snake_case_,"""word_embeddings""" ) elif layer_num == 1: trace.append("""position_embeddings""" ) _A : Optional[Any] = getattr(snake_case_,"""position_embeddings""" ) elif layer_num == 2: trace.append("""token_type_embeddings""" ) _A : Optional[Any] = getattr(snake_case_,"""token_type_embeddings""" ) else: raise ValueError(f'''Unknown embedding layer with name {full_name}''' ) trace.append("""weight""" ) _A : Any = getattr(snake_case_,"""weight""" ) elif m_name == "_attention_layer": # self-attention layer trace.extend(["""attention""", """self"""] ) _A : Optional[int] = getattr(snake_case_,"""attention""" ) _A : List[Any] = getattr(snake_case_,"""self""" ) elif m_name == "_attention_layer_norm": # output attention norm trace.extend(["""attention""", """output""", """LayerNorm"""] ) _A : Optional[int] = getattr(snake_case_,"""attention""" ) _A : Tuple = getattr(snake_case_,"""output""" ) _A : int = getattr(snake_case_,"""LayerNorm""" ) elif m_name == "_attention_output_dense": # output attention dense trace.extend(["""attention""", """output""", """dense"""] ) _A : Optional[Any] = getattr(snake_case_,"""attention""" ) _A : Any = getattr(snake_case_,"""output""" ) _A : Optional[int] = getattr(snake_case_,"""dense""" ) elif m_name == "_output_dense": # output dense trace.extend(["""output""", """dense"""] ) _A : Any = getattr(snake_case_,"""output""" ) _A : Union[str, Any] = getattr(snake_case_,"""dense""" ) elif m_name == "_output_layer_norm": # output dense trace.extend(["""output""", """LayerNorm"""] ) _A : str = getattr(snake_case_,"""output""" ) _A : Union[str, Any] = getattr(snake_case_,"""LayerNorm""" ) elif m_name == "_key_dense": # attention key trace.append("""key""" ) _A : Optional[int] = getattr(snake_case_,"""key""" ) elif m_name == "_query_dense": # attention query trace.append("""query""" ) _A : Optional[int] = getattr(snake_case_,"""query""" ) elif m_name == "_value_dense": # attention value trace.append("""value""" ) _A : Optional[Any] = getattr(snake_case_,"""value""" ) elif m_name == "_intermediate_dense": # attention intermediate dense trace.extend(["""intermediate""", """dense"""] ) _A : Dict = getattr(snake_case_,"""intermediate""" ) _A : Union[str, Any] = getattr(snake_case_,"""dense""" ) elif m_name == "_output_layer_norm": # output layer norm trace.append("""output""" ) _A : int = getattr(snake_case_,"""output""" ) # weights & biases elif m_name in ["bias", "beta"]: trace.append("""bias""" ) _A : List[str] = getattr(snake_case_,"""bias""" ) elif m_name in ["kernel", "gamma"]: trace.append("""weight""" ) _A : Optional[Any] = getattr(snake_case_,"""weight""" ) else: logger.warning(f'''Ignored {m_name}''' ) # for certain layers reshape is necessary _A : List[str] = """.""".join(snake_case_ ) if re.match(r"""(\S+)\.attention\.self\.(key|value|query)\.(bias|weight)""",snake_case_ ) or re.match( r"""(\S+)\.attention\.output\.dense\.weight""",snake_case_ ): _A : str = array.reshape(pointer.data.shape ) if "kernel" in full_name: _A : Tuple = array.transpose() if pointer.shape == array.shape: _A : Union[str, Any] = torch.from_numpy(snake_case_ ) else: raise ValueError( f'''Shape mismatch in layer {full_name}: Model expects shape {pointer.shape} but layer contains shape:''' f''' {array.shape}''' ) logger.info(f'''Successfully set variable {full_name} to PyTorch layer {trace}''' ) return model def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): # Instantiate model logger.info(f'''Loading model based on config from {config_path}...''' ) _A : Tuple = BertConfig.from_json_file(snake_case_ ) _A : Union[str, Any] = BertModel(snake_case_ ) # Load weights from checkpoint logger.info(f'''Loading weights from checkpoint {tf_checkpoint_path}...''' ) load_tfa_weights_in_bert(snake_case_,snake_case_,snake_case_ ) # Save pytorch-model logger.info(f'''Saving PyTorch model to {pytorch_dump_path}...''' ) torch.save(model.state_dict(),snake_case_ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( "--tf_checkpoint_path", type=str, required=True, help="Path to the TensorFlow 2.x checkpoint path." ) parser.add_argument( "--bert_config_file", type=str, required=True, help="The config json file corresponding to the BERT model. This specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", type=str, required=True, help="Path to the output PyTorch model (must include filename).", ) _snake_case = parser.parse_args() convert_tfa_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
54
def lowerCAmelCase_ ( snake_case_,snake_case_ ): _A : Union[str, Any] = """""" for i in table: res += inp[i - 1] return res def lowerCAmelCase_ ( snake_case_ ): return data[1:] + data[0] def lowerCAmelCase_ ( snake_case_,snake_case_ ): _A : Dict = """""" for i in range(len(snake_case_ ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def lowerCAmelCase_ ( snake_case_,snake_case_ ): _A : int = int("""0b""" + data[0] + data[-1],2 ) _A : Any = int("""0b""" + data[1:3],2 ) return bin(s[row][col] )[2:] def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_,snake_case_,snake_case_ ): _A : List[str] = message[:4] _A : List[Any] = message[4:] _A : Union[str, Any] = apply_table(snake_case_,snake_case_ ) _A : List[Any] = xor(snake_case_,snake_case_ ) _A : Optional[Any] = apply_sbox(snake_case_,temp[:4] ) # noqa: E741 _A : List[Any] = apply_sbox(snake_case_,temp[4:] ) _A : int = """0""" * (2 - len(snake_case_ )) + l # noqa: E741 _A : Union[str, Any] = """0""" * (2 - len(snake_case_ )) + r _A : List[Any] = apply_table(l + r,snake_case_ ) _A : Any = xor(snake_case_,snake_case_ ) return temp + right if __name__ == "__main__": _snake_case = input("Enter 10 bit key: ") _snake_case = input("Enter 8 bit message: ") _snake_case = [6, 3, 7, 4, 8, 5, 10, 9] _snake_case = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] _snake_case = [2, 4, 3, 1] _snake_case = [2, 6, 3, 1, 4, 8, 5, 7] _snake_case = [4, 1, 3, 5, 7, 2, 8, 6] _snake_case = [4, 1, 2, 3, 2, 3, 4, 1] _snake_case = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] _snake_case = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation _snake_case = apply_table(key, paa_table) _snake_case = temp[:5] _snake_case = temp[5:] _snake_case = left_shift(left) _snake_case = left_shift(right) _snake_case = apply_table(left + right, pa_table) _snake_case = left_shift(left) _snake_case = left_shift(right) _snake_case = left_shift(left) _snake_case = left_shift(right) _snake_case = apply_table(left + right, pa_table) # encryption _snake_case = apply_table(message, IP) _snake_case = function(expansion, sa, sa, keya, temp) _snake_case = temp[4:] + temp[:4] _snake_case = function(expansion, sa, sa, keya, temp) _snake_case = apply_table(temp, IP_inv) print("Cipher text is:", CT) # decryption _snake_case = apply_table(CT, IP) _snake_case = function(expansion, sa, sa, keya, temp) _snake_case = temp[4:] + temp[:4] _snake_case = function(expansion, sa, sa, keya, temp) _snake_case = apply_table(temp, IP_inv) print("Plain text after decypting is:", PT)
54
1
'''simple docstring''' from collections import deque def _A ( snake_case ) -> Any: _lowercase : Union[str, Any] = len(_UpperCamelCase ) _lowercase : Optional[Any] = deque() _lowercase : Dict = [False for _ in range(_UpperCamelCase )] _lowercase : str = [-1 for _ in range(_UpperCamelCase )] _lowercase : List[Any] = index_of[:] def strong_connect(snake_case , snake_case , snake_case ): _lowercase : Tuple = index # the number when this node is seen _lowercase : Any = index # lowest rank node reachable from here index += 1 stack.append(_UpperCamelCase ) _lowercase : Any = True for w in g[v]: if index_of[w] == -1: _lowercase : Optional[Any] = strong_connect(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) _lowercase : List[Any] = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) elif on_stack[w]: _lowercase : Dict = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) if lowlink_of[v] == index_of[v]: _lowercase : Any = [] _lowercase : List[str] = stack.pop() _lowercase : Any = False component.append(_UpperCamelCase ) while w != v: _lowercase : str = stack.pop() _lowercase : Union[str, Any] = False component.append(_UpperCamelCase ) components.append(_UpperCamelCase ) return index _lowercase : Any = [] for v in range(_UpperCamelCase ): if index_of[v] == -1: strong_connect(_UpperCamelCase , 0 , _UpperCamelCase ) return components def _A ( snake_case , snake_case ) -> Optional[int]: _lowercase : List[str] = [[] for _ in range(_UpperCamelCase )] for u, v in edges: g[u].append(_UpperCamelCase ) return g if __name__ == "__main__": # Test _snake_case = 7 _snake_case = [0, 0, 1, 2, 3, 3, 4, 4, 6] _snake_case = [1, 3, 2, 0, 1, 4, 5, 6, 5] _snake_case = [(u, v) for u, v in zip(source, target)] _snake_case = create_graph(n_vertices, edges) assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
245
'''simple docstring''' from math import factorial def __a ( _UpperCamelCase: int = 100 ) -> int: """simple docstring""" return sum(map(_UpperCamelCase , str(factorial(_UpperCamelCase ) ) ) ) if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
185
0
import argparse import ast import logging import os import sys import pandas as pd import torch from tqdm import tqdm from transformers import BartForConditionalGeneration, RagRetriever, RagSequenceForGeneration, RagTokenForGeneration from transformers import logging as transformers_logging sys.path.append(os.path.join(os.getcwd())) # noqa: E402 # isort:skip from utils_rag import exact_match_score, fa_score # noqa: E402 # isort:skip _snake_case : Union[str, Any] = logging.getLogger(__name__) logging.basicConfig(level=logging.INFO) transformers_logging.set_verbosity_info() def _A ( __snake_case :Dict ) -> Dict: """simple docstring""" if "token" in model_name_or_path: return "rag_token" if "sequence" in model_name_or_path: return "rag_sequence" if "bart" in model_name_or_path: return "bart" return None def _A ( __snake_case :int , __snake_case :Union[str, Any] , __snake_case :Tuple ) -> Optional[Any]: """simple docstring""" return max(metric_fn(__snake_case , __snake_case ) for gt in ground_truths ) def _A ( __snake_case :Optional[Any] , __snake_case :int , __snake_case :str ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = [line.strip() for line in open(__snake_case , "r" ).readlines()] __SCREAMING_SNAKE_CASE = [] if args.gold_data_mode == "qa": __SCREAMING_SNAKE_CASE = pd.read_csv(__snake_case , sep="\t" , header=__snake_case ) for answer_list in data[1]: __SCREAMING_SNAKE_CASE = ast.literal_eval(__snake_case ) answers.append(__snake_case ) else: __SCREAMING_SNAKE_CASE = [line.strip() for line in open(__snake_case , "r" ).readlines()] __SCREAMING_SNAKE_CASE = [[reference] for reference in references] __SCREAMING_SNAKE_CASE = __SCREAMING_SNAKE_CASE = __SCREAMING_SNAKE_CASE = 0 for prediction, ground_truths in zip(__snake_case , __snake_case ): total += 1 em += metric_max_over_ground_truths(__snake_case , __snake_case , __snake_case ) fa += metric_max_over_ground_truths(__snake_case , __snake_case , __snake_case ) __SCREAMING_SNAKE_CASE = 1_0_0.0 * em / total __SCREAMING_SNAKE_CASE = 1_0_0.0 * fa / total logger.info(f'''F1: {fa:.2f}''' ) logger.info(f'''EM: {em:.2f}''' ) def _A ( __snake_case :Tuple , __snake_case :int , __snake_case :Optional[Any] ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = args.k __SCREAMING_SNAKE_CASE = [line.strip() for line in open(__snake_case , "r" ).readlines()] __SCREAMING_SNAKE_CASE = [line.strip() for line in open(__snake_case , "r" ).readlines()] __SCREAMING_SNAKE_CASE = __SCREAMING_SNAKE_CASE = 0 for hypo, reference in zip(__snake_case , __snake_case ): __SCREAMING_SNAKE_CASE = set(hypo.split("\t" )[:k] ) __SCREAMING_SNAKE_CASE = set(reference.split("\t" ) ) total += 1 em += len(hypo_provenance & ref_provenance ) / k __SCREAMING_SNAKE_CASE = 1_0_0.0 * em / total logger.info(f'''Precision@{k}: {em: .2f}''' ) def _A ( __snake_case :Dict , __snake_case :int , __snake_case :List[str] ) -> int: """simple docstring""" def strip_title(__snake_case :int ): if title.startswith("\"" ): __SCREAMING_SNAKE_CASE = title[1:] if title.endswith("\"" ): __SCREAMING_SNAKE_CASE = title[:-1] return title __SCREAMING_SNAKE_CASE = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( __snake_case , return_tensors="pt" , padding=__snake_case , truncation=__snake_case , )["input_ids"].to(args.device ) __SCREAMING_SNAKE_CASE = rag_model.rag.question_encoder(__snake_case ) __SCREAMING_SNAKE_CASE = question_enc_outputs[0] __SCREAMING_SNAKE_CASE = rag_model.retriever( __snake_case , question_enc_pool_output.cpu().detach().to(torch.floataa ).numpy() , prefix=rag_model.rag.generator.config.prefix , n_docs=rag_model.config.n_docs , return_tensors="pt" , ) __SCREAMING_SNAKE_CASE = rag_model.retriever.index.get_doc_dicts(result.doc_ids ) __SCREAMING_SNAKE_CASE = [] for docs in all_docs: __SCREAMING_SNAKE_CASE = [strip_title(__snake_case ) for title in docs["title"]] provenance_strings.append("\t".join(__snake_case ) ) return provenance_strings def _A ( __snake_case :List[str] , __snake_case :Dict , __snake_case :Optional[int] ) -> str: """simple docstring""" with torch.no_grad(): __SCREAMING_SNAKE_CASE = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( __snake_case , return_tensors="pt" , padding=__snake_case , truncation=__snake_case ) __SCREAMING_SNAKE_CASE = inputs_dict.input_ids.to(args.device ) __SCREAMING_SNAKE_CASE = inputs_dict.attention_mask.to(args.device ) __SCREAMING_SNAKE_CASE = rag_model.generate( # rag_model overwrites generate __snake_case , attention_mask=__snake_case , num_beams=args.num_beams , min_length=args.min_length , max_length=args.max_length , early_stopping=__snake_case , num_return_sequences=1 , bad_words_ids=[[0, 0]] , ) __SCREAMING_SNAKE_CASE = rag_model.retriever.generator_tokenizer.batch_decode(__snake_case , skip_special_tokens=__snake_case ) if args.print_predictions: for q, a in zip(__snake_case , __snake_case ): logger.info("Q: {} - A: {}".format(__snake_case , __snake_case ) ) return answers def _A ( ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument( "--model_type" , choices=["rag_sequence", "rag_token", "bart"] , type=__snake_case , help=( "RAG model type: rag_sequence, rag_token or bart, if none specified, the type is inferred from the" " model_name_or_path" ) , ) parser.add_argument( "--index_name" , default=__snake_case , choices=["exact", "compressed", "legacy"] , type=__snake_case , help="RAG model retriever type" , ) parser.add_argument( "--index_path" , default=__snake_case , type=__snake_case , help="Path to the retrieval index" , ) parser.add_argument("--n_docs" , default=5 , type=__snake_case , help="Number of retrieved docs" ) parser.add_argument( "--model_name_or_path" , default=__snake_case , type=__snake_case , required=__snake_case , help="Path to pretrained checkpoints or model identifier from huggingface.co/models" , ) parser.add_argument( "--eval_mode" , choices=["e2e", "retrieval"] , default="e2e" , type=__snake_case , help=( "Evaluation mode, e2e calculates exact match and F1 of the downstream task, retrieval calculates" " precision@k." ) , ) parser.add_argument("--k" , default=1 , type=__snake_case , help="k for the precision@k calculation" ) parser.add_argument( "--evaluation_set" , default=__snake_case , type=__snake_case , required=__snake_case , help="Path to a file containing evaluation samples" , ) parser.add_argument( "--gold_data_path" , default=__snake_case , type=__snake_case , required=__snake_case , help="Path to a tab-separated file with gold samples" , ) parser.add_argument( "--gold_data_mode" , default="qa" , type=__snake_case , choices=["qa", "ans"] , help=( "Format of the gold data file" "qa - a single line in the following format: question [tab] answer_list" "ans - a single line of the gold file contains the expected answer string" ) , ) parser.add_argument( "--predictions_path" , type=__snake_case , default="predictions.txt" , help="Name of the predictions file, to be stored in the checkpoints directory" , ) parser.add_argument( "--eval_all_checkpoints" , action="store_true" , help="Evaluate all checkpoints starting with the same prefix as model_name ending and ending with step number" , ) parser.add_argument( "--eval_batch_size" , default=8 , type=__snake_case , help="Batch size per GPU/CPU for evaluation." , ) parser.add_argument( "--recalculate" , help="Recalculate predictions even if the prediction file exists" , action="store_true" , ) parser.add_argument( "--num_beams" , default=4 , type=__snake_case , help="Number of beams to be used when generating answers" , ) parser.add_argument("--min_length" , default=1 , type=__snake_case , help="Min length of the generated answers" ) parser.add_argument("--max_length" , default=50 , type=__snake_case , help="Max length of the generated answers" ) parser.add_argument( "--print_predictions" , action="store_true" , help="If True, prints predictions while evaluating." , ) parser.add_argument( "--print_docs" , action="store_true" , help="If True, prints docs retried while generating." , ) __SCREAMING_SNAKE_CASE = parser.parse_args() __SCREAMING_SNAKE_CASE = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) return args def _A ( __snake_case :List[Any] ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = {} if args.model_type is None: __SCREAMING_SNAKE_CASE = infer_model_type(args.model_name_or_path ) assert args.model_type is not None if args.model_type.startswith("rag" ): __SCREAMING_SNAKE_CASE = RagTokenForGeneration if args.model_type == "rag_token" else RagSequenceForGeneration __SCREAMING_SNAKE_CASE = args.n_docs if args.index_name is not None: __SCREAMING_SNAKE_CASE = args.index_name if args.index_path is not None: __SCREAMING_SNAKE_CASE = args.index_path else: __SCREAMING_SNAKE_CASE = BartForConditionalGeneration __SCREAMING_SNAKE_CASE = ( [f.path for f in os.scandir(args.model_name_or_path ) if f.is_dir()] if args.eval_all_checkpoints else [args.model_name_or_path] ) logger.info("Evaluate the following checkpoints: %s" , __snake_case ) __SCREAMING_SNAKE_CASE = get_scores if args.eval_mode == "e2e" else get_precision_at_k __SCREAMING_SNAKE_CASE = evaluate_batch_eae if args.eval_mode == "e2e" else evaluate_batch_retrieval for checkpoint in checkpoints: if os.path.exists(args.predictions_path ) and (not args.recalculate): logger.info("Calculating metrics based on an existing predictions file: {}".format(args.predictions_path ) ) score_fn(__snake_case , args.predictions_path , args.gold_data_path ) continue logger.info("***** Running evaluation for {} *****".format(__snake_case ) ) logger.info(" Batch size = %d" , args.eval_batch_size ) logger.info(" Predictions will be stored under {}".format(args.predictions_path ) ) if args.model_type.startswith("rag" ): __SCREAMING_SNAKE_CASE = RagRetriever.from_pretrained(__snake_case , **__snake_case ) __SCREAMING_SNAKE_CASE = model_class.from_pretrained(__snake_case , retriever=__snake_case , **__snake_case ) model.retriever.init_retrieval() else: __SCREAMING_SNAKE_CASE = model_class.from_pretrained(__snake_case , **__snake_case ) model.to(args.device ) with open(args.evaluation_set , "r" ) as eval_file, open(args.predictions_path , "w" ) as preds_file: __SCREAMING_SNAKE_CASE = [] for line in tqdm(__snake_case ): questions.append(line.strip() ) if len(__snake_case ) == args.eval_batch_size: __SCREAMING_SNAKE_CASE = evaluate_batch_fn(__snake_case , __snake_case , __snake_case ) preds_file.write("\n".join(__snake_case ) + "\n" ) preds_file.flush() __SCREAMING_SNAKE_CASE = [] if len(__snake_case ) > 0: __SCREAMING_SNAKE_CASE = evaluate_batch_fn(__snake_case , __snake_case , __snake_case ) preds_file.write("\n".join(__snake_case ) ) preds_file.flush() score_fn(__snake_case , args.predictions_path , args.gold_data_path ) if __name__ == "__main__": _snake_case : Dict = get_args() main(args)
214
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _snake_case : Optional[int] = {'configuration_fnet': ['FNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FNetConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : str = ['FNetTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Dict = ['FNetTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ 'FNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'FNetForMaskedLM', 'FNetForMultipleChoice', 'FNetForNextSentencePrediction', 'FNetForPreTraining', 'FNetForQuestionAnswering', 'FNetForSequenceClassification', 'FNetForTokenClassification', 'FNetLayer', 'FNetModel', 'FNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_fnet import FNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet import FNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_fnet_fast import FNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_fnet import ( FNET_PRETRAINED_MODEL_ARCHIVE_LIST, FNetForMaskedLM, FNetForMultipleChoice, FNetForNextSentencePrediction, FNetForPreTraining, FNetForQuestionAnswering, FNetForSequenceClassification, FNetForTokenClassification, FNetLayer, FNetModel, FNetPreTrainedModel, ) else: import sys _snake_case : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
214
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ = logging.get_logger(__name__) snake_case_ = { 'uclanlp/visualbert-vqa': 'https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json', 'uclanlp/visualbert-vqa-pre': 'https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json', 'uclanlp/visualbert-vqa-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json' ), 'uclanlp/visualbert-vcr': 'https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json', 'uclanlp/visualbert-vcr-pre': 'https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json', 'uclanlp/visualbert-vcr-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json' ), 'uclanlp/visualbert-nlvr2': 'https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json', 'uclanlp/visualbert-nlvr2-pre': 'https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json', 'uclanlp/visualbert-nlvr2-coco-pre': ( 'https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json' ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): _A = "visual_bert" def __init__( self , lowercase__=3_0522 , lowercase__=768 , lowercase__=512 , lowercase__=12 , lowercase__=12 , lowercase__=3072 , lowercase__="gelu" , lowercase__=0.1 , lowercase__=0.1 , lowercase__=512 , lowercase__=2 , lowercase__=0.02 , lowercase__=1e-12 , lowercase__=False , lowercase__=True , lowercase__=1 , lowercase__=0 , lowercase__=2 , **lowercase__ , ): """simple docstring""" super().__init__(pad_token_id=lowercase__ , bos_token_id=lowercase__ , eos_token_id=lowercase__ , **lowercase__ ) SCREAMING_SNAKE_CASE_ : str = vocab_size SCREAMING_SNAKE_CASE_ : List[str] = max_position_embeddings SCREAMING_SNAKE_CASE_ : int = hidden_size SCREAMING_SNAKE_CASE_ : Optional[Any] = visual_embedding_dim SCREAMING_SNAKE_CASE_ : Union[str, Any] = num_hidden_layers SCREAMING_SNAKE_CASE_ : str = num_attention_heads SCREAMING_SNAKE_CASE_ : Any = intermediate_size SCREAMING_SNAKE_CASE_ : Dict = hidden_act SCREAMING_SNAKE_CASE_ : List[str] = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : List[Any] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : Optional[Any] = initializer_range SCREAMING_SNAKE_CASE_ : List[Any] = type_vocab_size SCREAMING_SNAKE_CASE_ : Dict = layer_norm_eps SCREAMING_SNAKE_CASE_ : Optional[Any] = bypass_transformer SCREAMING_SNAKE_CASE_ : str = special_visual_initialize
421
'''simple docstring''' import re def __lowerCamelCase ( SCREAMING_SNAKE_CASE_ : str ) -> bool: """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = re.compile(R"^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$" ) if match := re.search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator('+918827897895'))
421
1
"""simple docstring""" import copy import os import cva import numpy as np from matplotlib import pyplot as plt class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self ): '''simple docstring''' _UpperCamelCase : Any = "" _UpperCamelCase : List[Any] = "" _UpperCamelCase : int = [] _UpperCamelCase : str = 0 _UpperCamelCase : Optional[Any] = 2_56 _UpperCamelCase : Union[str, Any] = 0 _UpperCamelCase : Tuple = 0 _UpperCamelCase : Dict = 0 _UpperCamelCase : Tuple = 0 def lowercase_ (self , lowerCAmelCase__ ): '''simple docstring''' _UpperCamelCase : Optional[int] = cva.imread(lowerCAmelCase__ , 0 ) _UpperCamelCase : Optional[int] = copy.deepcopy(self.img ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : str = plt.hist(self.img.ravel() , 2_56 , [0, 2_56] , label="x" ) _UpperCamelCase : Union[str, Any] = np.sum(lowerCAmelCase__ ) for i in range(len(lowerCAmelCase__ ) ): _UpperCamelCase : str = x[i] / self.k self.sk += prk _UpperCamelCase : Dict = (self.L - 1) * self.sk if self.rem != 0: _UpperCamelCase : int = int(last % last ) _UpperCamelCase : int = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(lowerCAmelCase__ ) _UpperCamelCase : Optional[Any] = int(np.ma.count(self.img ) / self.img[1].size ) _UpperCamelCase : List[str] = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): _UpperCamelCase : Any = self.img[j][i] if num != self.last_list[num]: _UpperCamelCase : Any = self.last_list[num] cva.imwrite("output_data/output.jpg" , self.img ) def lowercase_ (self ): '''simple docstring''' plt.hist(self.img.ravel() , 2_56 , [0, 2_56] ) def lowercase_ (self ): '''simple docstring''' cva.imshow("Output-Image" , self.img ) cva.imshow("Input-Image" , self.original_image ) cva.waitKey(50_00 ) cva.destroyAllWindows() if __name__ == "__main__": _SCREAMING_SNAKE_CASE = os.path.join(os.path.basename(__file__), """image_data/input.jpg""") _SCREAMING_SNAKE_CASE = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
239
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { """facebook/data2vec-text-base""": """https://huggingface.co/data2vec/resolve/main/config.json""", } class _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __UpperCAmelCase = """data2vec-text""" def __init__(self , lowerCAmelCase__=3_05_22 , lowerCAmelCase__=7_68 , lowerCAmelCase__=12 , lowerCAmelCase__=12 , lowerCAmelCase__=30_72 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=5_12 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=1 , lowerCAmelCase__=0 , lowerCAmelCase__=2 , lowerCAmelCase__="absolute" , lowerCAmelCase__=True , lowerCAmelCase__=None , **lowerCAmelCase__ , ): '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) _UpperCamelCase : Optional[Any] = vocab_size _UpperCamelCase : Dict = hidden_size _UpperCamelCase : List[Any] = num_hidden_layers _UpperCamelCase : Tuple = num_attention_heads _UpperCamelCase : Any = hidden_act _UpperCamelCase : Any = intermediate_size _UpperCamelCase : List[str] = hidden_dropout_prob _UpperCamelCase : Tuple = attention_probs_dropout_prob _UpperCamelCase : Dict = max_position_embeddings _UpperCamelCase : List[Any] = type_vocab_size _UpperCamelCase : List[Any] = initializer_range _UpperCamelCase : Dict = layer_norm_eps _UpperCamelCase : Any = position_embedding_type _UpperCamelCase : List[Any] = use_cache _UpperCamelCase : str = classifier_dropout class _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' @property def lowercase_ (self ): '''simple docstring''' if self.task == "multiple-choice": _UpperCamelCase : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: _UpperCamelCase : Tuple = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
239
1
from typing import TYPE_CHECKING from ....utils import _LazyModule UpperCAmelCase : Optional[Any] = {"""tokenization_tapex""": ["""TapexTokenizer"""]} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys UpperCAmelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
563
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase__ =get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowerCAmelCase__( __lowercase , unittest.TestCase ): '''simple docstring''' __snake_case = XLNetTokenizer __snake_case = XLNetTokenizerFast __snake_case = True __snake_case = True def UpperCamelCase_ ( self ) -> Optional[Any]: super().setUp() # We have a SentencePiece fixture for testing _SCREAMING_SNAKE_CASE : List[Any] = XLNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> Optional[int]: _SCREAMING_SNAKE_CASE : Union[str, Any] = "<s>" _SCREAMING_SNAKE_CASE : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def UpperCamelCase_ ( self ) -> List[Any]: _SCREAMING_SNAKE_CASE : List[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "<eod>" ) self.assertEqual(len(__lowerCamelCase ) , 1_0_0_6 ) def UpperCamelCase_ ( self ) -> str: self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_0 ) def UpperCamelCase_ ( self ) -> Tuple: _SCREAMING_SNAKE_CASE : List[Any] = XLNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowerCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2] ) _SCREAMING_SNAKE_CASE : str = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 0, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 0, 4] ) _SCREAMING_SNAKE_CASE : int = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def UpperCamelCase_ ( self ) -> str: _SCREAMING_SNAKE_CASE : List[Any] = XLNetTokenizer(__lowerCamelCase , do_lower_case=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "", "i", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "se", ".", ] , ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["▁he", "ll", "o"] ) def UpperCamelCase_ ( self ) -> Optional[int]: _SCREAMING_SNAKE_CASE : List[Any] = XLNetTokenizer(__lowerCamelCase , do_lower_case=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "se", ".", ] , ) @slow def UpperCamelCase_ ( self ) -> List[str]: _SCREAMING_SNAKE_CASE : Union[str, Any] = XLNetTokenizer.from_pretrained("xlnet-base-cased" ) _SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.encode("sequence builders" , add_special_tokens=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.encode("multi-sequence build" , add_special_tokens=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[str] = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Any = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase , __lowerCamelCase ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def UpperCamelCase_ ( self ) -> int: # fmt: off _SCREAMING_SNAKE_CASE : List[Any] = {"input_ids": [[1_7, 2_1_4_4_2, 2_7_0, 1_7, 1_0, 1_4_6_4_5, 3_1_8, 3_4, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 7_7_5_2, 2_2_0_1_8, 2_3, 2_1, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 3_3_5_2, 1_4_4_3_1, 1_3, 5_5_0_0, 1_1, 1_1_7_6, 5_8_0, 1_3, 1_6_8_1_9, 4_7_9_7, 2_3, 1_7, 1_0, 1_7_1_3_5, 6_5_8, 1_9, 4_5_7, 7_9_3_2, 1_3, 1_8_4, 1_9, 3_1_5_4, 1_7_1_3_5, 6_4_6_8, 1_9, 1_4_0_4, 1_2_2_6_9, 1_9, 4_2_2_9, 5_3_5_6, 1_6_2_6_4, 4_6, 1_9, 1_7, 2_0_5_4_5, 1_0_3_9_5, 9, 9, 9, 1_1, 2_8, 6_4_2_1, 9_5_3_1, 2_0_7_2_9, 1_7, 1_0, 3_5_3, 1_7_0_2_2, 1_1, 2_1, 6_4_2_1, 9_5_3_1, 1_6_9_4_9, 1_7, 1_0, 1_1_5_0_9, 7_5_3, 1_1, 3_3, 9_5, 2_4_2_1, 7_3_8_5, 9_5_6, 1_4_4_3_1, 2_6_2_6, 2_5, 8_4_2, 7_3_8_5, 4_8_3_6, 2_1, 1_4_2_9, 2_2_7_2, 9_8_5_5, 3_1_2_0, 1_6_1, 2_4_7_3_8, 1_9, 1_3_2_0_3, 6_5_8, 2_1_8, 7_8_7, 2_1, 4_3_0, 1_8_4_8_2, 8_4_7, 2_6_3_7, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2_2, 2_2_1_7_8, 2_7, 1_0_6_4, 2_2, 9_5_6, 1_3, 1_1_1_0_1, 1_4_2_9, 5_8_5_4, 2_4_3_1_3, 1_8_9_5_3, 4_0, 4_2_2, 2_4_3_6_6, 6_8, 1_7_5_8, 3_7, 1_0_4_8_3, 1_4_2_5_7, 3_1, 2_0_7, 2_6_3, 2_1, 2_0_3, 3_7_7_3, 2_5, 7_1, 9_7_3_5, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2, 2_0_4_9, 3_4_4_2, 1_7, 1_3_8_9_4, 3_3_8_0, 2_3, 9_5, 1_8, 1_7_6_3_4, 2_2_8_8, 9, 4, 3]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name="xlnet-base-cased" , revision="c841166438c31ec7ca9a106dee7bb312b73ae511" , )
249
0
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np UpperCamelCase : int = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 UpperCamelCase : str = typing.Union[np.floataa, int, float] # noqa: UP007 def UpperCamelCase_ ( __a , __a ) -> VectorOut: return np.sqrt(np.sum((np.asarray(__a ) - np.asarray(__a )) ** 2 ) ) def UpperCamelCase_ ( __a , __a ) -> VectorOut: return sum((va - va) ** 2 for va, va in zip(__a , __a ) ) ** (1 / 2) if __name__ == "__main__": def UpperCamelCase_ ( ) -> None: from timeit import timeit print("Without Numpy" ) print( timeit( "euclidean_distance_no_np([1, 2, 3], [4, 5, 6])" , number=10_000 , globals=globals() , ) ) print("With Numpy" ) print( timeit( "euclidean_distance([1, 2, 3], [4, 5, 6])" , number=10_000 , globals=globals() , ) ) benchmark()
703
def UpperCamelCase_ ( __a , __a ) -> float: if density <= 0: raise ValueError("Impossible fluid density" ) if bulk_modulus <= 0: raise ValueError("Impossible bulk modulus" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
151
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=3 , lowerCamelCase__=1_8 , lowerCamelCase__=3_0 , lowerCamelCase__=4_0_0 , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = image_size _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size_divisor _lowerCamelCase = do_rescale def snake_case__ ( self ): return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = GLPNImageProcessor if is_vision_available() else None def snake_case__ ( self ): _lowerCamelCase = GLPNImageProcessingTester(self ) @property def snake_case__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self ): _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''size_divisor''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''resample''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_rescale''' ) ) def snake_case__ ( self ): pass def snake_case__ ( self ): # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def snake_case__ ( self ): # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def snake_case__ ( self ): # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
661
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __SCREAMING_SNAKE_CASE : Tuple = { '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } __SCREAMING_SNAKE_CASE : str = {'''facebook/blenderbot-3B''': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase_( ) -> str: _lowerCamelCase = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) _lowerCamelCase = bs[:] _lowerCamelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(lowercase_ ) cs.append(2**8 + n ) n += 1 _lowerCamelCase = [chr(lowercase_ ) for n in cs] return dict(zip(lowercase_ , lowercase_ ) ) def lowerCAmelCase_( lowercase_ : str ) -> Dict: _lowerCamelCase = set() _lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _lowerCamelCase = char return pairs class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : int = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Tuple = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__="replace" , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<mask>" , lowerCamelCase__=False , **lowerCamelCase__ , ): _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else bos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else eos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else sep_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else cls_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else unk_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ , **lowerCamelCase__ , ) with open(lowerCamelCase__ , encoding='''utf-8''' ) as vocab_handle: _lowerCamelCase = json.load(lowerCamelCase__ ) _lowerCamelCase = {v: k for k, v in self.encoder.items()} _lowerCamelCase = errors # how to handle errors in decoding _lowerCamelCase = bytes_to_unicode() _lowerCamelCase = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ , encoding='''utf-8''' ) as merges_handle: _lowerCamelCase = merges_handle.read().split('''\n''' )[1:-1] _lowerCamelCase = [tuple(merge.split() ) for merge in bpe_merges] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = {} _lowerCamelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _lowerCamelCase = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def snake_case__ ( self ): return len(self.encoder ) def snake_case__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def snake_case__ ( self , lowerCamelCase__ ): if token in self.cache: return self.cache[token] _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _lowerCamelCase = min(lowerCamelCase__ , key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _lowerCamelCase , _lowerCamelCase = bigram _lowerCamelCase = [] _lowerCamelCase = 0 while i < len(lowerCamelCase__ ): try: _lowerCamelCase = word.index(lowerCamelCase__ , lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _lowerCamelCase = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = new_word if len(lowerCamelCase__ ) == 1: break else: _lowerCamelCase = get_pairs(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = word return word def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for token in re.findall(self.pat , lowerCamelCase__ ): _lowerCamelCase = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(''' ''' ) ) return bpe_tokens def snake_case__ ( self , lowerCamelCase__ ): return self.encoder.get(lowerCamelCase__ , self.encoder.get(self.unk_token ) ) def snake_case__ ( self , lowerCamelCase__ ): return self.decoder.get(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = ''''''.join(lowerCamelCase__ ) _lowerCamelCase = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase__ , ensure_ascii=lowerCamelCase__ ) + '''\n''' ) _lowerCamelCase = 0 with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) _lowerCamelCase = token_index writer.write(''' '''.join(lowerCamelCase__ ) + '''\n''' ) index += 1 return vocab_file, merge_file def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , **lowerCamelCase__ ): _lowerCamelCase = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _lowerCamelCase = ''' ''' + text return (text, kwargs) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): return token_ids_a + [self.eos_token_id] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(''' ''' + text ) else: # Generated responses should contain them already. inputs.append(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = self.encode(lowerCamelCase__ ) if len(lowerCamelCase__ ) > self.model_max_length: _lowerCamelCase = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
661
1
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss __lowerCamelCase : Optional[Any] = pytest.mark.integration @require_faiss class UpperCAmelCase ( _lowercase ): def UpperCAmelCase__ (self : Union[str, Any] ) -> Tuple: lowercase = Dataset.from_dict({"filename": ["my_name-train" + "_" + str(A__ ) for x in np.arange(3_0 ).tolist()]} ) return dset def UpperCAmelCase__ (self : List[Any] ) -> Optional[Any]: import faiss lowercase = self._create_dummy_dataset() lowercase = dset.map( lambda A__ , A__ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=A__ , keep_in_memory=A__ ) lowercase = dset.add_faiss_index("vecs" , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT ) lowercase , lowercase = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) dset.drop_index("vecs" ) def UpperCAmelCase__ (self : Optional[int] ) -> Any: import faiss lowercase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name="vecs" , batch_size=1_0_0 , metric_type=faiss.METRIC_INNER_PRODUCT , ) lowercase , lowercase = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def UpperCAmelCase__ (self : Tuple ) -> Dict: import faiss lowercase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name="vecs" , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=A__ ) as tmp_file: dset.save_faiss_index("vecs" , tmp_file.name ) dset.load_faiss_index("vecs2" , tmp_file.name ) os.unlink(tmp_file.name ) lowercase , lowercase = dset.get_nearest_examples("vecs2" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def UpperCAmelCase__ (self : Any ) -> int: lowercase = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((3_0, 5) ) * np.arange(3_0 ).reshape(-1 , 1 ) , index_name="vecs" ) dset.drop_index("vecs" ) self.assertRaises(A__ , partial(dset.get_nearest_examples , "vecs2" , np.ones(5 , dtype=np.floataa ) ) ) def UpperCAmelCase__ (self : Optional[Any] ) -> Dict: from elasticsearch import Elasticsearch lowercase = self._create_dummy_dataset() with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: lowercase = {"acknowledged": True} mocked_bulk.return_value([(True, None)] * 3_0 ) lowercase = {"hits": {"hits": [{"_score": 1, "_id": 2_9}]}} lowercase = Elasticsearch() dset.add_elasticsearch_index("filename" , es_client=A__ ) lowercase , lowercase = dset.get_nearest_examples("filename" , "my_name-train_29" ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) @require_faiss class UpperCAmelCase ( _lowercase ): def UpperCAmelCase__ (self : Optional[Any] ) -> Tuple: import faiss lowercase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 1_0 ) # single query lowercase = np.zeros(5 , dtype=np.floataa ) lowercase = 1 lowercase , lowercase = index.search(A__ ) self.assertRaises(A__ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries lowercase = np.eye(5 , dtype=np.floataa )[::-1] lowercase , lowercase = index.search_batch(A__ ) self.assertRaises(A__ , index.search_batch , queries[0] ) lowercase = [scores[0] for scores in total_scores] lowercase = [indices[0] for indices in total_indices] self.assertGreater(np.min(A__ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , A__ ) def UpperCAmelCase__ (self : Tuple ) -> Optional[Any]: import faiss lowercase = FaissIndex(string_factory="Flat" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) lowercase = FaissIndex(string_factory="LSH" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(A__ ): lowercase = FaissIndex(string_factory="Flat" , custom_index=faiss.IndexFlat(5 ) ) def UpperCAmelCase__ (self : Tuple ) -> Union[str, Any]: import faiss lowercase = faiss.IndexFlat(5 ) lowercase = FaissIndex(custom_index=A__ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def UpperCAmelCase__ (self : List[Any] ) -> List[Any]: import faiss lowercase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=A__ ) as tmp_file: index.save(tmp_file.name ) lowercase = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) lowercase = np.zeros(5 , dtype=np.floataa ) lowercase = 1 lowercase , lowercase = index.search(A__ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def UpperCAmelCase_ ( lowerCAmelCase_ ): """simple docstring""" import faiss lowercase = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) lowercase = "index.faiss" lowercase = f'mock://{index_name}' index.save(lowerCAmelCase_ , storage_options=mockfs.storage_options ) lowercase = FaissIndex.load(lowerCAmelCase_ , storage_options=mockfs.storage_options ) lowercase = np.zeros(5 , dtype=np.floataa ) lowercase = 1 lowercase , lowercase = index.search(lowerCAmelCase_ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class UpperCAmelCase ( _lowercase ): def UpperCAmelCase__ (self : Dict ) -> Union[str, Any]: from elasticsearch import Elasticsearch with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: lowercase = Elasticsearch() lowercase = {"acknowledged": True} lowercase = ElasticSearchIndex(es_client=A__ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(["foo", "bar", "foobar"] ) # single query lowercase = "foo" lowercase = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} lowercase , lowercase = index.search(A__ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout lowercase = "foo" lowercase = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} lowercase , lowercase = index.search(A__ , request_timeout=3_0 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries lowercase = ["foo", "bar", "foobar"] lowercase = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} lowercase , lowercase = index.search_batch(A__ ) lowercase = [scores[0] for scores in total_scores] lowercase = [indices[0] for indices in total_indices] self.assertGreater(np.min(A__ ) , 0 ) self.assertListEqual([1, 1, 1] , A__ ) # batched queries with timeout lowercase = ["foo", "bar", "foobar"] lowercase = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} lowercase , lowercase = index.search_batch(A__ , request_timeout=3_0 ) lowercase = [scores[0] for scores in total_scores] lowercase = [indices[0] for indices in total_indices] self.assertGreater(np.min(A__ ) , 0 ) self.assertListEqual([1, 1, 1] , A__ )
459
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): UpperCAmelCase : Optional[int] = ViTImageProcessor if is_vision_available() else None @property def UpperCAmelCase__ (self : Optional[int] ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ (self : str ) -> int: lowercase = (3, 3_2, 1_2_8) lowercase = tempfile.mkdtemp() # fmt: off lowercase = ["[GO]", "[s]", "0", "1", "2", "3", "4", "5", "6", "7", "8", "9", "a", "b", "c", "d", "e", "f", "g", "h", "i", "j", "k", "l", "m", "n", "o", "p", "q", "r", "s", "t", "u", "v", "w", "x", "y", "z"] # fmt: on lowercase = dict(zip(A__ , range(len(A__ ) ) ) ) lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(A__ ) + "\n" ) lowercase = { "do_normalize": False, "do_resize": True, "image_processor_type": "ViTImageProcessor", "resample": 3, "size": {"height": 3_2, "width": 1_2_8}, } lowercase = os.path.join(self.tmpdirname , A__ ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(A__ , A__ ) def UpperCAmelCase__ (self : Dict , **A__ : List[Any] ) -> Union[str, Any]: return MgpstrTokenizer.from_pretrained(self.tmpdirname , **A__ ) def UpperCAmelCase__ (self : Any , **A__ : Tuple ) -> Tuple: return ViTImageProcessor.from_pretrained(self.tmpdirname , **A__ ) def UpperCAmelCase__ (self : Optional[int] ) -> int: shutil.rmtree(self.tmpdirname ) def UpperCAmelCase__ (self : str ) -> List[str]: lowercase = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) lowercase = Image.fromarray(np.moveaxis(A__ , 0 , -1 ) ) return image_input def UpperCAmelCase__ (self : str ) -> List[Any]: lowercase = self.get_tokenizer() lowercase = self.get_image_processor() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) processor.save_pretrained(self.tmpdirname ) lowercase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=A__ ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A__ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , A__ ) def UpperCAmelCase__ (self : Any ) -> Optional[int]: lowercase = self.get_tokenizer() lowercase = self.get_image_processor() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) processor.save_pretrained(self.tmpdirname ) lowercase = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) lowercase = self.get_image_processor(do_normalize=A__ , padding_value=1.0 ) lowercase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=A__ , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , A__ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , A__ ) def UpperCAmelCase__ (self : List[Any] ) -> List[Any]: lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) lowercase = self.prepare_image_inputs() lowercase = image_processor(A__ , return_tensors="np" ) lowercase = processor(images=A__ , return_tensors="np" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def UpperCAmelCase__ (self : List[Any] ) -> Union[str, Any]: lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) lowercase = "test" lowercase = processor(text=A__ ) lowercase = tokenizer(A__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCAmelCase__ (self : int ) -> Optional[int]: lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) lowercase = "test" lowercase = self.prepare_image_inputs() lowercase = processor(text=A__ , images=A__ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "labels"] ) # test if it raises when no input is passed with pytest.raises(A__ ): processor() def UpperCAmelCase__ (self : List[Any] ) -> Union[str, Any]: lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) lowercase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] lowercase = processor.char_decode(A__ ) lowercase = tokenizer.batch_decode(A__ ) lowercase = [seq.replace(" " , "" ) for seq in decoded_tok] self.assertListEqual(A__ , A__ ) def UpperCAmelCase__ (self : Any ) -> Optional[int]: lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) lowercase = None lowercase = self.prepare_image_inputs() lowercase = processor(text=A__ , images=A__ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def UpperCAmelCase__ (self : int ) -> Optional[Any]: lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = MgpstrProcessor(tokenizer=A__ , image_processor=A__ ) lowercase = torch.randn(1 , 2_7 , 3_8 ) lowercase = torch.randn(1 , 2_7 , 5_0_2_5_7 ) lowercase = torch.randn(1 , 2_7 , 3_0_5_2_2 ) lowercase = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ["generated_text", "scores", "char_preds", "bpe_preds", "wp_preds"] )
459
1
'''simple docstring''' import math import tensorflow as tf from packaging import version def _UpperCamelCase ( __UpperCamelCase ) -> Any: lowerCamelCase_ = tf.convert_to_tensor(__UpperCamelCase ) lowerCamelCase_ = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) ,x.dtype ) )) return x * cdf def _UpperCamelCase ( __UpperCamelCase ) -> Optional[int]: lowerCamelCase_ = tf.convert_to_tensor(__UpperCamelCase ) lowerCamelCase_ = tf.cast(math.pi ,x.dtype ) lowerCamelCase_ = tf.cast(0.04_4715 ,x.dtype ) lowerCamelCase_ = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(__UpperCamelCase ,3 )) )) return x * cdf def _UpperCamelCase ( __UpperCamelCase ) -> Optional[Any]: lowerCamelCase_ = tf.convert_to_tensor(__UpperCamelCase ) return x * tf.tanh(tf.math.softplus(__UpperCamelCase ) ) def _UpperCamelCase ( __UpperCamelCase ) -> str: lowerCamelCase_ = tf.convert_to_tensor(__UpperCamelCase ) lowerCamelCase_ = tf.cast(0.04_4715 ,x.dtype ) lowerCamelCase_ = tf.cast(0.79_7884_5608 ,x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def _UpperCamelCase ( __UpperCamelCase ) -> str: lowerCamelCase_ = tf.convert_to_tensor(__UpperCamelCase ) lowerCamelCase_ = tf.cast(1.702 ,x.dtype ) return x * tf.math.sigmoid(coeff * x ) def _UpperCamelCase ( __UpperCamelCase ) -> int: return tf.clip_by_value(_gelu(__UpperCamelCase ) ,-10 ,10 ) def _UpperCamelCase ( __UpperCamelCase ,__UpperCamelCase=-1 ) -> List[Any]: lowerCamelCase_ ,lowerCamelCase_ = tf.split(__UpperCamelCase ,2 ,axis=__UpperCamelCase ) return a * tf.math.sigmoid(__UpperCamelCase ) if version.parse(tf.version.VERSION) >= version.parse("2.4"): def _UpperCamelCase ( __UpperCamelCase ) -> Dict: return tf.keras.activations.gelu(__UpperCamelCase ,approximate=__UpperCamelCase ) A_ = tf.keras.activations.gelu A_ = approximate_gelu_wrap else: A_ = _gelu A_ = _gelu_new A_ = { "gelu": gelu, "gelu_10": gelu_aa, "gelu_fast": gelu_fast, "gelu_new": gelu_new, "glu": glu, "mish": mish, "quick_gelu": quick_gelu, "relu": tf.keras.activations.relu, "sigmoid": tf.keras.activations.sigmoid, "silu": tf.keras.activations.swish, "swish": tf.keras.activations.swish, "tanh": tf.keras.activations.tanh, } def _UpperCamelCase ( __UpperCamelCase ) -> Tuple: if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(f'''function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}''' )
42
'''simple docstring''' import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A_ = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A_ = 250_004 A_ = 250_020 @require_sentencepiece @require_tokenizers class UpperCAmelCase ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ = MBartTokenizer SCREAMING_SNAKE_CASE_ = MBartTokenizerFast SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = True def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowerCamelCase_ = MBartTokenizer(SCREAMING_SNAKE_CASE_ , keep_accents=SCREAMING_SNAKE_CASE_ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' lowerCamelCase_ = MBartTokenizer(SCREAMING_SNAKE_CASE_ , keep_accents=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer.tokenize('This is a test' ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowerCamelCase_ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( SCREAMING_SNAKE_CASE_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowerCamelCase_ = tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE_ ) self.assertListEqual( SCREAMING_SNAKE_CASE_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) lowerCamelCase_ = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE_ ) self.assertListEqual( SCREAMING_SNAKE_CASE_ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def UpperCamelCase( self ) -> int: '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return lowerCamelCase_ = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-random-mbart', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCamelCase_ = self.rust_tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = self.tokenizer_class.from_pretrained(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.save_pretrained(SCREAMING_SNAKE_CASE_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) lowerCamelCase_ = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) # Save tokenizer rust, legacy_format=True lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(SCREAMING_SNAKE_CASE_ , legacy_format=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.save_pretrained(SCREAMING_SNAKE_CASE_ ) # Checks it save with the same files self.assertSequenceEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) # Save tokenizer rust, legacy_format=False lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = tokenizer_r.save_pretrained(SCREAMING_SNAKE_CASE_ , legacy_format=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.save_pretrained(SCREAMING_SNAKE_CASE_ ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCamelCase_ = tokenizer_r.from_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = tokenizer_p.from_pretrained(SCREAMING_SNAKE_CASE_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ = 'facebook/mbart-large-en-ro' SCREAMING_SNAKE_CASE_ = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.', ] SCREAMING_SNAKE_CASE_ = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei' ' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor' ' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] SCREAMING_SNAKE_CASE_ = [82_74, 12_78_73, 2_59_16, 7, 86_22, 20_71, 4_38, 6_74_85, 53, 18_78_95, 23, 5_17_12, 2, EN_CODE] @classmethod def UpperCamelCase( cls ) -> Optional[Any]: '''simple docstring''' lowerCamelCase_ = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang='en_XX' , tgt_lang='ro_RO' ) lowerCamelCase_ = 1 return cls def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ar_AR'] , 250001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['en_EN'] , 250004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['ro_RO'] , 250020 ) def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , SCREAMING_SNAKE_CASE_ ) def UpperCamelCase( self ) -> Optional[int]: '''simple docstring''' self.assertIn(SCREAMING_SNAKE_CASE_ , self.tokenizer.all_special_ids ) lowerCamelCase_ = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] lowerCamelCase_ = self.tokenizer.decode(SCREAMING_SNAKE_CASE_ , skip_special_tokens=SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertNotIn(self.tokenizer.eos_token , SCREAMING_SNAKE_CASE_ ) def UpperCamelCase( self ) -> List[Any]: '''simple docstring''' lowerCamelCase_ = ['this is gunna be a long sentence ' * 20] assert isinstance(src_text[0] , SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = 10 lowerCamelCase_ = self.tokenizer(SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , SCREAMING_SNAKE_CASE_ ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) def UpperCamelCase( self ) -> Tuple: '''simple docstring''' self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['<mask>', 'ar_AR'] ) , [250026, 250001] ) def UpperCamelCase( self ) -> List[str]: '''simple docstring''' lowerCamelCase_ = tempfile.mkdtemp() lowerCamelCase_ = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = MBartTokenizer.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , SCREAMING_SNAKE_CASE_ ) @require_torch def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=SCREAMING_SNAKE_CASE_ , return_tensors='pt' ) lowerCamelCase_ = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def UpperCamelCase( self ) -> Dict: '''simple docstring''' lowerCamelCase_ = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=len(self.expected_src_tokens ) , return_tensors='pt' , ) lowerCamelCase_ = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) lowerCamelCase_ = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , SCREAMING_SNAKE_CASE_ ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE] ) def UpperCamelCase( self ) -> Optional[Any]: '''simple docstring''' lowerCamelCase_ = self.tokenizer(self.src_text , padding=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=3 , return_tensors='pt' ) lowerCamelCase_ = self.tokenizer( text_target=self.tgt_text , padding=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=10 , return_tensors='pt' ) lowerCamelCase_ = targets['input_ids'] lowerCamelCase_ = shift_tokens_right(SCREAMING_SNAKE_CASE_ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def UpperCamelCase( self ) -> Any: '''simple docstring''' lowerCamelCase_ = self.tokenizer._build_translation_inputs( 'A test' , return_tensors='pt' , src_lang='en_XX' , tgt_lang='ar_AR' ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ) , { # A, test, EOS, en_XX 'input_ids': [[62, 3034, 2, 250004]], 'attention_mask': [[1, 1, 1, 1]], # ar_AR 'forced_bos_token_id': 250001, } , )
42
1
import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __A : Optional[int] = logging.get_logger(__name__) __A : str = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } __A : Union[str, Any] = { "vocab_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json"}, "merges_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt"}, "tokenizer_config_file": { "facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json" }, } __A : List[Any] = {"facebook/blenderbot-3B": 128} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def UpperCAmelCase ( ) -> Any: '''simple docstring''' __lowerCAmelCase = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) __lowerCAmelCase = bs[:] __lowerCAmelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCamelCase__ ) cs.append(2**8 + n ) n += 1 __lowerCAmelCase = [chr(UpperCamelCase__ ) for n in cs] return dict(zip(UpperCamelCase__ , UpperCamelCase__ ) ) def UpperCAmelCase ( UpperCamelCase__ ) -> str: '''simple docstring''' __lowerCAmelCase = set() __lowerCAmelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowerCAmelCase = char return pairs class lowercase_ ( lowerCAmelCase__ ): __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = ["input_ids", "attention_mask"] def __init__( self: Optional[Any], _lowercase: Any, _lowercase: Optional[Any], _lowercase: Optional[Any]="replace", _lowercase: Tuple="<s>", _lowercase: Tuple="</s>", _lowercase: Union[str, Any]="</s>", _lowercase: Optional[Any]="<s>", _lowercase: Optional[Any]="<unk>", _lowercase: str="<pad>", _lowercase: Dict="<mask>", _lowercase: Any=False, **_lowercase: List[Any], ): '''simple docstring''' __lowerCAmelCase = AddedToken(_lowercase, lstrip=_lowercase, rstrip=_lowercase) if isinstance(_lowercase, _lowercase) else bos_token __lowerCAmelCase = AddedToken(_lowercase, lstrip=_lowercase, rstrip=_lowercase) if isinstance(_lowercase, _lowercase) else eos_token __lowerCAmelCase = AddedToken(_lowercase, lstrip=_lowercase, rstrip=_lowercase) if isinstance(_lowercase, _lowercase) else sep_token __lowerCAmelCase = AddedToken(_lowercase, lstrip=_lowercase, rstrip=_lowercase) if isinstance(_lowercase, _lowercase) else cls_token __lowerCAmelCase = AddedToken(_lowercase, lstrip=_lowercase, rstrip=_lowercase) if isinstance(_lowercase, _lowercase) else unk_token __lowerCAmelCase = AddedToken(_lowercase, lstrip=_lowercase, rstrip=_lowercase) if isinstance(_lowercase, _lowercase) else pad_token # Mask token behave like a normal word, i.e. include the space before it __lowerCAmelCase = AddedToken(_lowercase, lstrip=_lowercase, rstrip=_lowercase) if isinstance(_lowercase, _lowercase) else mask_token super().__init__( errors=_lowercase, bos_token=_lowercase, eos_token=_lowercase, unk_token=_lowercase, sep_token=_lowercase, cls_token=_lowercase, pad_token=_lowercase, mask_token=_lowercase, add_prefix_space=_lowercase, **_lowercase, ) with open(_lowercase, encoding="""utf-8""") as vocab_handle: __lowerCAmelCase = json.load(_lowercase) __lowerCAmelCase = {v: k for k, v in self.encoder.items()} __lowerCAmelCase = errors # how to handle errors in decoding __lowerCAmelCase = bytes_to_unicode() __lowerCAmelCase = {v: k for k, v in self.byte_encoder.items()} with open(_lowercase, encoding="""utf-8""") as merges_handle: __lowerCAmelCase = merges_handle.read().split("""\n""")[1:-1] __lowerCAmelCase = [tuple(merge.split()) for merge in bpe_merges] __lowerCAmelCase = dict(zip(_lowercase, range(len(_lowercase)))) __lowerCAmelCase = {} __lowerCAmelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions __lowerCAmelCase = re.compile(R"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""") @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def _lowercase ( self: Optional[Any]): '''simple docstring''' return len(self.encoder) def _lowercase ( self: Dict): '''simple docstring''' return dict(self.encoder, **self.added_tokens_encoder) def _lowercase ( self: Dict, _lowercase: Dict): '''simple docstring''' if token in self.cache: return self.cache[token] __lowerCAmelCase = tuple(_lowercase) __lowerCAmelCase = get_pairs(_lowercase) if not pairs: return token while True: __lowerCAmelCase = min(_lowercase, key=lambda _lowercase: self.bpe_ranks.get(_lowercase, float("""inf"""))) if bigram not in self.bpe_ranks: break __lowerCAmelCase , __lowerCAmelCase = bigram __lowerCAmelCase = [] __lowerCAmelCase = 0 while i < len(_lowercase): try: __lowerCAmelCase = word.index(_lowercase, _lowercase) except ValueError: new_word.extend(word[i:]) break else: new_word.extend(word[i:j]) __lowerCAmelCase = j if word[i] == first and i < len(_lowercase) - 1 and word[i + 1] == second: new_word.append(first + second) i += 2 else: new_word.append(word[i]) i += 1 __lowerCAmelCase = tuple(_lowercase) __lowerCAmelCase = new_word if len(_lowercase) == 1: break else: __lowerCAmelCase = get_pairs(_lowercase) __lowerCAmelCase = """ """.join(_lowercase) __lowerCAmelCase = word return word def _lowercase ( self: Any, _lowercase: int): '''simple docstring''' __lowerCAmelCase = [] for token in re.findall(self.pat, _lowercase): __lowerCAmelCase = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""")) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_lowercase).split(""" """)) return bpe_tokens def _lowercase ( self: Union[str, Any], _lowercase: Dict): '''simple docstring''' return self.encoder.get(_lowercase, self.encoder.get(self.unk_token)) def _lowercase ( self: Optional[int], _lowercase: Any): '''simple docstring''' return self.decoder.get(_lowercase) def _lowercase ( self: Dict, _lowercase: Tuple): '''simple docstring''' __lowerCAmelCase = """""".join(_lowercase) __lowerCAmelCase = bytearray([self.byte_decoder[c] for c in text]).decode("""utf-8""", errors=self.errors) return text def _lowercase ( self: Dict, _lowercase: str, _lowercase: Optional[str] = None): '''simple docstring''' if not os.path.isdir(_lowercase): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''') return __lowerCAmelCase = os.path.join( _lowercase, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""]) __lowerCAmelCase = os.path.join( _lowercase, (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""]) with open(_lowercase, """w""", encoding="""utf-8""") as f: f.write(json.dumps(self.encoder, indent=2, sort_keys=_lowercase, ensure_ascii=_lowercase) + """\n""") __lowerCAmelCase = 0 with open(_lowercase, """w""", encoding="""utf-8""") as writer: writer.write("""#version: 0.2\n""") for bpe_tokens, token_index in sorted(self.bpe_ranks.items(), key=lambda _lowercase: kv[1]): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""") __lowerCAmelCase = token_index writer.write(""" """.join(_lowercase) + """\n""") index += 1 return vocab_file, merge_file def _lowercase ( self: Dict, _lowercase: List[int], _lowercase: Optional[List[int]] = None, _lowercase: bool = False): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowercase, token_ids_a=_lowercase, already_has_special_tokens=_lowercase) if token_ids_a is None: return [1] + ([0] * len(_lowercase)) + [1] return [1] + ([0] * len(_lowercase)) + [1, 1] + ([0] * len(_lowercase)) + [1] def _lowercase ( self: List[str], _lowercase: List[int], _lowercase: Optional[List[int]] = None): '''simple docstring''' __lowerCAmelCase = [self.sep_token_id] __lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def _lowercase ( self: Dict, _lowercase: List[Any], _lowercase: List[str]=False, **_lowercase: Dict): '''simple docstring''' __lowerCAmelCase = kwargs.pop("""add_prefix_space""", self.add_prefix_space) if (is_split_into_words or add_prefix_space) and (len(_lowercase) > 0 and not text[0].isspace()): __lowerCAmelCase = """ """ + text return (text, kwargs) def _lowercase ( self: Any, _lowercase: List[int], _lowercase: Optional[List[int]] = None): '''simple docstring''' return token_ids_a + [self.eos_token_id] def _lowercase ( self: Optional[Any], _lowercase: "Conversation"): '''simple docstring''' __lowerCAmelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(""" """ + text) else: # Generated responses should contain them already. inputs.append(_lowercase) __lowerCAmelCase = """ """.join(_lowercase) __lowerCAmelCase = self.encode(_lowercase) if len(_lowercase) > self.model_max_length: __lowerCAmelCase = input_ids[-self.model_max_length :] logger.warning(f'''Trimmed input from conversation as it was longer than {self.model_max_length} tokens.''') return input_ids
334
def UpperCAmelCase ( UpperCamelCase__ ) -> str: '''simple docstring''' if isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError("""'float' object cannot be interpreted as an integer""" ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError("""'str' object cannot be interpreted as an integer""" ) if num == 0: return "0b0" __lowerCAmelCase = False if num < 0: __lowerCAmelCase = True __lowerCAmelCase = -num __lowerCAmelCase = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(UpperCamelCase__ ) for e in binary ) return "0b" + "".join(str(UpperCamelCase__ ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
334
1
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' UpperCAmelCase = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] UpperCAmelCase = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } UpperCAmelCase = F"""{src_lang}-{tgt_lang}""" UpperCAmelCase = F""" --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR's WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = \"facebook/wmt19-{src_lang}-{tgt_lang}\" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = \"{texts[src_lang]}\" input_ids = tokenizer.encode(input, return_tensors=\"pt\") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn't seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn't support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR's WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) """ os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) UpperCAmelCase = os.path.join(UpperCamelCase__ , '''README.md''' ) print(F"""Generating {path}""" ) with open(UpperCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(UpperCamelCase__ ) # make sure we are under the root of the project __A : Optional[Any] = Path(__file__).resolve().parent.parent.parent __A : Optional[Any] = repo_dir / "model_cards" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __A , __A , __A : List[str] = model_name.split("-") __A : int = model_cards_dir / "facebook" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
130
def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = 0 , UpperCamelCase__ = 0 ) -> int: '''simple docstring''' UpperCAmelCase = right or len(UpperCamelCase__ ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(UpperCamelCase__ , UpperCamelCase__ , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
130
1
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' snake_case_ : List[Any] = (1 + 2_4 * n) ** 0.5 return ((1 + root) / 6) % 1 == 0 def __lowerCAmelCase ( __UpperCamelCase : int = 5_0_0_0 ): '''simple docstring''' snake_case_ : str = [(i * (3 * i - 1)) // 2 for i in range(1 , __UpperCamelCase )] for i, pentagonal_i in enumerate(__UpperCamelCase ): for j in range(__UpperCamelCase , len(__UpperCamelCase ) ): snake_case_ : Optional[int] = pentagonal_nums[j] snake_case_ : List[Any] = pentagonal_i + pentagonal_j snake_case_ : Optional[Any] = pentagonal_j - pentagonal_i if is_pentagonal(__UpperCamelCase ) and is_pentagonal(__UpperCamelCase ): return b return -1 if __name__ == "__main__": print(F'''{solution() = }''')
715
"""simple docstring""" from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' def is_in_circle(__UpperCamelCase : float , __UpperCamelCase : float ) -> bool: snake_case_ : Dict = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle snake_case_ : Tuple = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(__UpperCamelCase ) ) # The ratio of the area for circle to square is pi/4. snake_case_ : Union[str, Any] = proportion * 4 print(F'The estimated value of pi is {pi_estimate}' ) print(F'The numpy value of pi is {pi}' ) print(F'The total error is {abs(pi - pi_estimate )}' ) def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Callable[[float], float] , __UpperCamelCase : float = 0.0 , __UpperCamelCase : float = 1.0 , ): '''simple docstring''' return mean( function_to_integrate(uniform(__UpperCamelCase , __UpperCamelCase ) ) for _ in range(__UpperCamelCase ) ) * (max_value - min_value) def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : float = 0.0 , __UpperCamelCase : float = 1.0 ): '''simple docstring''' def identity_function(__UpperCamelCase : float ) -> float: return x snake_case_ : int = area_under_curve_estimator( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : str = (max_value * max_value - min_value * min_value) / 2 print("""******************""" ) print(F'Estimating area under y=x where x varies from {min_value} to {max_value}' ) print(F'Estimated value is {estimated_value}' ) print(F'Expected value is {expected_value}' ) print(F'Total error is {abs(estimated_value - expected_value )}' ) print("""******************""" ) def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' def function_to_integrate(__UpperCamelCase : float ) -> float: return sqrt(4.0 - x * x ) snake_case_ : List[Any] = area_under_curve_estimator( __UpperCamelCase , __UpperCamelCase , 0.0 , 2.0 ) print("""******************""" ) print("""Estimating pi using area_under_curve_estimator""" ) print(F'Estimated value is {estimated_value}' ) print(F'Expected value is {pi}' ) print(F'Total error is {abs(estimated_value - pi )}' ) print("""******************""" ) if __name__ == "__main__": import doctest doctest.testmod()
21
0
import unittest from transformers import AutoConfig, AutoTokenizer, BertConfig, TensorType, is_flax_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, slow if is_flax_available(): import jax from transformers.models.auto.modeling_flax_auto import FlaxAutoModel from transformers.models.bert.modeling_flax_bert import FlaxBertModel from transformers.models.roberta.modeling_flax_roberta import FlaxRobertaModel @require_flax class lowerCamelCase ( unittest.TestCase ): @slow def A( self): for model_name in ["bert-base-cased", "bert-large-uncased"]: with self.subTest(lowercase__): __UpperCAmelCase : Dict = AutoConfig.from_pretrained(lowercase__) self.assertIsNotNone(lowercase__) self.assertIsInstance(lowercase__ , lowercase__) __UpperCAmelCase : str = FlaxAutoModel.from_pretrained(lowercase__) self.assertIsNotNone(lowercase__) self.assertIsInstance(lowercase__ , lowercase__) @slow def A( self): for model_name in ["roberta-base", "roberta-large"]: with self.subTest(lowercase__): __UpperCAmelCase : Optional[int] = AutoConfig.from_pretrained(lowercase__) self.assertIsNotNone(lowercase__) self.assertIsInstance(lowercase__ , lowercase__) __UpperCAmelCase : str = FlaxAutoModel.from_pretrained(lowercase__) self.assertIsNotNone(lowercase__) self.assertIsInstance(lowercase__ , lowercase__) @slow def A( self): for model_name in ["bert-base-cased", "bert-large-uncased"]: __UpperCAmelCase : Any = AutoTokenizer.from_pretrained(lowercase__) __UpperCAmelCase : Any = FlaxBertModel.from_pretrained(lowercase__) __UpperCAmelCase : List[str] = tokenizer('''Do you support jax jitted function?''' , return_tensors=TensorType.JAX) @jax.jit def eval(**lowercase__): return model(**lowercase__) eval(**lowercase__).block_until_ready() @slow def A( self): for model_name in ["roberta-base", "roberta-large"]: __UpperCAmelCase : int = AutoTokenizer.from_pretrained(lowercase__) __UpperCAmelCase : int = FlaxRobertaModel.from_pretrained(lowercase__) __UpperCAmelCase : Union[str, Any] = tokenizer('''Do you support jax jitted function?''' , return_tensors=TensorType.JAX) @jax.jit def eval(**lowercase__): return model(**lowercase__) eval(**lowercase__).block_until_ready() def A( self): with self.assertRaisesRegex( lowercase__ , '''bert-base is not a local folder and is not a valid model identifier'''): __UpperCAmelCase : Optional[int] = FlaxAutoModel.from_pretrained('''bert-base''') def A( self): with self.assertRaisesRegex( lowercase__ , r'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)'''): __UpperCAmelCase : Union[str, Any] = FlaxAutoModel.from_pretrained(lowercase__ , revision='''aaaaaa''') def A( self): with self.assertRaisesRegex( lowercase__ , '''hf-internal-testing/config-no-model does not appear to have a file named flax_model.msgpack''' , ): __UpperCAmelCase : Optional[int] = FlaxAutoModel.from_pretrained('''hf-internal-testing/config-no-model''') def A( self): with self.assertRaisesRegex(lowercase__ , '''Use `from_pt=True` to load this model'''): __UpperCAmelCase : Dict = FlaxAutoModel.from_pretrained('''hf-internal-testing/tiny-bert-pt-only''')
462
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class lowerCamelCase ( _UpperCamelCase ): _lowerCAmelCase : torch.FloatTensor _lowerCAmelCase : torch.FloatTensor _lowerCAmelCase : Optional[torch.FloatTensor] = None class lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): _lowerCAmelCase : Tuple = 2 @register_to_config def __init__( self , lowercase__ = 0.0_2 , lowercase__ = 1_0_0 , lowercase__ = 1.0_0_7 , lowercase__ = 8_0 , lowercase__ = 0.0_5 , lowercase__ = 5_0 , ): # standard deviation of the initial noise distribution __UpperCAmelCase : Union[str, Any] = sigma_max # setable values __UpperCAmelCase : int = None __UpperCAmelCase : np.IntTensor = None __UpperCAmelCase : torch.FloatTensor = None # sigma(t_i) def A( self , lowercase__ , lowercase__ = None): return sample def A( self , lowercase__ , lowercase__ = None): __UpperCAmelCase : int = num_inference_steps __UpperCAmelCase : List[str] = np.arange(0 , self.num_inference_steps)[::-1].copy() __UpperCAmelCase : Any = torch.from_numpy(lowercase__).to(lowercase__) __UpperCAmelCase : Union[str, Any] = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __UpperCAmelCase : Tuple = torch.tensor(lowercase__ , dtype=torch.floataa , device=lowercase__) def A( self , lowercase__ , lowercase__ , lowercase__ = None): if self.config.s_min <= sigma <= self.config.s_max: __UpperCAmelCase : int = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1) else: __UpperCAmelCase : int = 0 # sample eps ~ N(0, S_noise^2 * I) __UpperCAmelCase : List[str] = self.config.s_noise * randn_tensor(sample.shape , generator=lowercase__).to(sample.device) __UpperCAmelCase : Optional[int] = sigma + gamma * sigma __UpperCAmelCase : List[Any] = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def A( self , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = True , ): __UpperCAmelCase : str = sample_hat + sigma_hat * model_output __UpperCAmelCase : Tuple = (sample_hat - pred_original_sample) / sigma_hat __UpperCAmelCase : str = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowercase__ , derivative=lowercase__ , pred_original_sample=lowercase__) def A( self , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = True , ): __UpperCAmelCase : Any = sample_prev + sigma_prev * model_output __UpperCAmelCase : List[str] = (sample_prev - pred_original_sample) / sigma_prev __UpperCAmelCase : Union[str, Any] = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=lowercase__ , derivative=lowercase__ , pred_original_sample=lowercase__) def A( self , lowercase__ , lowercase__ , lowercase__): raise NotImplementedError()
462
1
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers snake_case_ = """3""" print("""Python version:""", sys.version) print("""transformers version:""", transformers.__version__) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) print("""NCCL version:""", torch.cuda.nccl.version()) except ImportError: print("""Torch version:""", None) try: import deepspeed print("""DeepSpeed version:""", deepspeed.__version__) except ImportError: print("""DeepSpeed version:""", None) try: import tensorflow as tf print("""TensorFlow version:""", tf.__version__) print("""TF GPUs available:""", bool(tf.config.list_physical_devices("""GPU"""))) print("""Number of TF GPUs available:""", len(tf.config.list_physical_devices("""GPU"""))) except ImportError: print("""TensorFlow version:""", None)
718
'''simple docstring''' from __future__ import annotations class _lowercase : def __init__( self , _UpperCAmelCase ): A : str = data A : Node | None = None A : Node | None = None def _lowerCamelCase( UpperCamelCase__ : Node | None ) -> None: # In Order traversal of the tree if tree: display(tree.left ) print(tree.data ) display(tree.right ) def _lowerCamelCase( UpperCamelCase__ : Node | None ) -> int: return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def _lowerCamelCase( UpperCamelCase__ : Node ) -> bool: if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def _lowerCamelCase( ) -> None: # Main function for testing. A : Optional[int] = Node(1 ) A : Tuple = Node(2 ) A : Dict = Node(3 ) A : List[str] = Node(4 ) A : Union[str, Any] = Node(5 ) A : str = Node(6 ) A : Any = Node(7 ) A : str = Node(8 ) A : Optional[int] = Node(9 ) print(is_full_binary_tree(UpperCamelCase__ ) ) print(depth_of_tree(UpperCamelCase__ ) ) print('''Tree is: ''' ) display(UpperCamelCase__ ) if __name__ == "__main__": main()
537
0
'''simple docstring''' import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values __a: Tuple = argparse.ArgumentParser() parser.add_argument("""--user""", type=str, default="""ubuntu""") parser.add_argument("""--host""", type=str, default="""localhost""") parser.add_argument("""--key_path""", type=str, default=None) parser.add_argument("""--instance""", type=str, default="""V100:1""") parser.add_argument("""--provider""", type=str, default="""cheapest""") parser.add_argument("""--use_spot""", type=bool, default=False) parser.add_argument("""--example""", type=str, default="""pytorch/text-generation/run_generation.py""") __a: List[str] = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError("""Cannot specify both BYO and on-demand cluster args""") __a: Dict = rh.cluster( name="""rh-cluster""", ips=[args.host], ssh_creds={"""ssh_user""": args.user, """ssh_private_key""": args.key_path} ) else: __a: Any = rh.cluster( name="""rh-cluster""", instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) __a: int = args.example.rsplit("""/""", 1)[0] # Set up remote environment cluster.install_packages(["""pip:./"""]) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F'pip install -r transformers/examples/{example_dir}/requirements.txt']) cluster.run(["""pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117"""]) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F'python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}']) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
152
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class UpperCAmelCase : '''simple docstring''' def snake_case__ ( self : List[Any] ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) snake_case_ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) snake_case_ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) snake_case_ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=__lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) snake_case_ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def snake_case__ ( self : List[str] ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = TaEncoderModel.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) snake_case_ = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-t5" ) torch.manual_seed(0 ) snake_case_ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ "ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D", ] , mid_block_type="UNetMidBlock2DSimpleCrossAttn" , up_block_types=["SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="text" , addition_embed_type_num_heads=2 , cross_attention_norm="group_norm" , resnet_time_scale_shift="scale_shift" , act_fn="gelu" , class_embed_type="timestep" , mid_block_scale_factor=1.414 , time_embedding_act_fn="gelu" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) snake_case_ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , thresholding=__lowercase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type="epsilon" , variance_type="learned_range" , ) torch.manual_seed(0 ) snake_case_ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule="squaredcos_cap_v2" , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) snake_case_ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def snake_case__ ( self : List[str] ): """simple docstring""" snake_case_ = self.get_dummy_components() snake_case_ = self.pipeline_class(**__lowercase ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) snake_case_ = self.get_dummy_inputs(__lowercase ) snake_case_ = inputs["prompt"] snake_case_ = inputs["generator"] snake_case_ = inputs["num_inference_steps"] snake_case_ = inputs["output_type"] if "image" in inputs: snake_case_ = inputs["image"] else: snake_case_ = None if "mask_image" in inputs: snake_case_ = inputs["mask_image"] else: snake_case_ = None if "original_image" in inputs: snake_case_ = inputs["original_image"] else: snake_case_ = None snake_case_ , snake_case_ = pipe.encode_prompt(__lowercase ) # inputs with prompt converted to embeddings snake_case_ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: snake_case_ = image if mask_image is not None: snake_case_ = mask_image if original_image is not None: snake_case_ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(__lowercase , __lowercase , __lowercase ) snake_case_ = pipe(**__lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__lowercase ) snake_case_ = self.pipeline_class.from_pretrained(__lowercase ) pipe_loaded.to(__lowercase ) pipe_loaded.set_progress_bar_config(disable=__lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(__lowercase , __lowercase ) is None , f"`{optional_component}` did not stay set to None after loading." , ) snake_case_ = self.get_dummy_inputs(__lowercase ) snake_case_ = inputs["generator"] snake_case_ = inputs["num_inference_steps"] snake_case_ = inputs["output_type"] # inputs with prompt converted to embeddings snake_case_ = { "prompt_embeds": prompt_embeds, "negative_prompt_embeds": negative_prompt_embeds, "generator": generator, "num_inference_steps": num_inference_steps, "output_type": output_type, } if image is not None: snake_case_ = image if mask_image is not None: snake_case_ = mask_image if original_image is not None: snake_case_ = original_image snake_case_ = pipe_loaded(**__lowercase )[0] snake_case_ = np.abs(to_np(__lowercase ) - to_np(__lowercase ) ).max() self.assertLess(__lowercase , 1E-4 ) def snake_case__ ( self : Union[str, Any] ): """simple docstring""" snake_case_ = self.get_dummy_components() snake_case_ = self.pipeline_class(**__lowercase ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) snake_case_ = self.get_dummy_inputs(__lowercase ) snake_case_ = pipe(**__lowercase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__lowercase ) snake_case_ = self.pipeline_class.from_pretrained(__lowercase ) pipe_loaded.to(__lowercase ) pipe_loaded.set_progress_bar_config(disable=__lowercase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests snake_case_ = self.get_dummy_inputs(__lowercase ) snake_case_ = pipe_loaded(**__lowercase )[0] snake_case_ = np.abs(to_np(__lowercase ) - to_np(__lowercase ) ).max() self.assertLess(__lowercase , 1E-4 )
376
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { """hustvl/yolos-small""": """https://huggingface.co/hustvl/yolos-small/resolve/main/config.json""", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" A__ : int = "yolos" def __init__( self , SCREAMING_SNAKE_CASE__=768 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=3072 , SCREAMING_SNAKE_CASE__="gelu" , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=0.0_2 , SCREAMING_SNAKE_CASE__=1e-12 , SCREAMING_SNAKE_CASE__=[512, 864] , SCREAMING_SNAKE_CASE__=16 , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=100 , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=1 , SCREAMING_SNAKE_CASE__=5 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=5 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=0.1 , **SCREAMING_SNAKE_CASE__ , ) -> int: super().__init__(**SCREAMING_SNAKE_CASE__ ) A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = initializer_range A__ = layer_norm_eps A__ = image_size A__ = patch_size A__ = num_channels A__ = qkv_bias A__ = num_detection_tokens A__ = use_mid_position_embeddings A__ = auxiliary_loss # Hungarian matcher A__ = class_cost A__ = bbox_cost A__ = giou_cost # Loss coefficients A__ = bbox_loss_coefficient A__ = giou_loss_coefficient A__ = eos_coefficient class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" A__ : Union[str, Any] = version.parse("1.11" ) @property def snake_case__ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def snake_case__ ( self ) -> float: return 1e-4 @property def snake_case__ ( self ) -> int: return 12
562
"""simple docstring""" import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {name: getattr(transformers, name + """Fast""") for name in SLOW_TO_FAST_CONVERTERS} def _lowerCamelCase ( UpperCAmelCase_ : Optional[Any], UpperCAmelCase_ : Any, UpperCAmelCase_ : List[Any], UpperCAmelCase_ : List[str] ) -> Tuple: """simple docstring""" if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(F"""Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.""" ) if tokenizer_name is None: A__ = TOKENIZER_CLASSES else: A__ = {tokenizer_name: getattr(UpperCAmelCase_, tokenizer_name + "Fast" )} logger.info(F"""Loading tokenizer classes: {tokenizer_names}""" ) for tokenizer_name in tokenizer_names: A__ = TOKENIZER_CLASSES[tokenizer_name] A__ = True if checkpoint_name is None: A__ = list(tokenizer_class.max_model_input_sizes.keys() ) else: A__ = [checkpoint_name] logger.info(F"""For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}""" ) for checkpoint in checkpoint_names: logger.info(F"""Loading {tokenizer_class.__class__.__name__} {checkpoint}""" ) # Load tokenizer A__ = tokenizer_class.from_pretrained(UpperCAmelCase_, force_download=UpperCAmelCase_ ) # Save fast tokenizer logger.info(F"""Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}""" ) # For organization names we create sub-directories if "/" in checkpoint: A__ , A__ = checkpoint.split("/" ) A__ = os.path.join(UpperCAmelCase_, UpperCAmelCase_ ) elif add_prefix: A__ = checkpoint A__ = dump_path else: A__ = None A__ = dump_path logger.info(F"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: A__ = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] A__ = file_path.split(UpperCAmelCase_ )[-1][0] if next_char == "/": A__ = os.path.join(UpperCAmelCase_, UpperCAmelCase_ ) A__ = None logger.info(F"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) A__ = tokenizer.save_pretrained( UpperCAmelCase_, legacy_format=UpperCAmelCase_, filename_prefix=UpperCAmelCase_ ) logger.info(F"""=> File names {file_names}""" ) for file_name in file_names: if not file_name.endswith("tokenizer.json" ): os.remove(UpperCAmelCase_ ) logger.info(F"""=> removing {file_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--dump_path""", default=None, type=str, required=True, help="""Path to output generated fast tokenizer files.""" ) parser.add_argument( """--tokenizer_name""", default=None, type=str, help=( f'Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will ' """download and convert all the checkpoints from AWS.""" ), ) parser.add_argument( """--checkpoint_name""", default=None, type=str, help="""Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.""", ) parser.add_argument( """--force_download""", action="""store_true""", help="""Re-download checkpoints.""", ) UpperCamelCase = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
562
1
"""simple docstring""" import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup __UpperCAmelCase = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36' ' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582' } def lowerCAmelCase ( __UpperCamelCase = "dhaka" , __UpperCamelCase = 5 ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = min(__UpperCamelCase , 50 ) # Prevent abuse! UpperCAmelCase__ : int = { """q""": query, """tbm""": """isch""", """hl""": """en""", """ijn""": """0""", } UpperCAmelCase__ : Dict = requests.get("""https://www.google.com/search""" , params=__UpperCamelCase , headers=__UpperCamelCase ) UpperCAmelCase__ : Any = BeautifulSoup(html.text , """html.parser""" ) UpperCAmelCase__ : List[Any] = """""".join( re.findall(r"""AF_initDataCallback\(([^<]+)\);""" , str(soup.select("""script""" ) ) ) ) UpperCAmelCase__ : Dict = json.dumps(__UpperCamelCase ) UpperCAmelCase__ : List[str] = json.loads(__UpperCamelCase ) UpperCAmelCase__ : List[str] = re.findall( r"""\[\"GRID_STATE0\",null,\[\[1,\[0,\".*?\",(.*),\"All\",""" , __UpperCamelCase , ) if not matched_google_image_data: return 0 UpperCAmelCase__ : Tuple = re.sub( r"""\[\"(https\:\/\/encrypted-tbn0\.gstatic\.com\/images\?.*?)\",\d+,\d+\]""" , """""" , str(__UpperCamelCase ) , ) UpperCAmelCase__ : str = re.findall( r"""(?:\'|,),\[\"(https:|http.*?)\",\d+,\d+\]""" , __UpperCamelCase , ) for index, fixed_full_res_image in enumerate(__UpperCamelCase ): if index >= max_images: return index UpperCAmelCase__ : Dict = bytes(__UpperCamelCase , """ascii""" ).decode( """unicode-escape""" ) UpperCAmelCase__ : Optional[Any] = bytes(__UpperCamelCase , """ascii""" ).decode( """unicode-escape""" ) UpperCAmelCase__ : Optional[Any] = urllib.request.build_opener() UpperCAmelCase__ : List[Any] = [ ( """User-Agent""", """Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36""" """ (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582""", ) ] urllib.request.install_opener(__UpperCamelCase ) UpperCAmelCase__ : Tuple = F"query_{query.replace(' ' , '_' )}" if not os.path.exists(__UpperCamelCase ): os.makedirs(__UpperCamelCase ) urllib.request.urlretrieve( # noqa: S310 __UpperCamelCase , F"{path_name}/original_size_img_{index}.jpg" ) return index if __name__ == "__main__": try: __UpperCAmelCase = download_images_from_google_query(sys.argv[1]) print(F"{image_count} images were downloaded to disk.") except IndexError: print('Please provide a search term.') raise
65
"""simple docstring""" import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class lowercase: def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=1_3 , __SCREAMING_SNAKE_CASE=7 , __SCREAMING_SNAKE_CASE=True , __SCREAMING_SNAKE_CASE=True , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=True , __SCREAMING_SNAKE_CASE=9_9 , __SCREAMING_SNAKE_CASE=3_2 , __SCREAMING_SNAKE_CASE=5 , __SCREAMING_SNAKE_CASE=4 , __SCREAMING_SNAKE_CASE=3_7 , __SCREAMING_SNAKE_CASE="gelu" , __SCREAMING_SNAKE_CASE=0.1 , __SCREAMING_SNAKE_CASE=0.1 , __SCREAMING_SNAKE_CASE=5_1_2 , __SCREAMING_SNAKE_CASE=1_6 , __SCREAMING_SNAKE_CASE=2 , __SCREAMING_SNAKE_CASE=0.02 , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=4 , __SCREAMING_SNAKE_CASE=None , ) -> int: """simple docstring""" a__ = parent a__ = batch_size a__ = seq_length a__ = is_training a__ = use_input_mask a__ = use_token_type_ids a__ = use_labels a__ = vocab_size a__ = hidden_size a__ = num_hidden_layers a__ = num_attention_heads a__ = intermediate_size a__ = hidden_act a__ = hidden_dropout_prob a__ = attention_probs_dropout_prob a__ = max_position_embeddings a__ = type_vocab_size a__ = type_sequence_label_size a__ = initializer_range a__ = num_labels a__ = num_choices a__ = scope def lowercase__ ( self ) -> Optional[Any]: """simple docstring""" a__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a__ = None if self.use_input_mask: a__ = random_attention_mask([self.batch_size, self.seq_length] ) a__ = None if self.use_token_type_ids: a__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a__ = None a__ = None a__ = None if self.use_labels: a__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a__ = ids_tensor([self.batch_size] , self.num_choices ) a__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase__ ( self ) -> Any: """simple docstring""" return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) def lowercase__ ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" a__ = LlamaModel(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() a__ = model(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ) a__ = model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , ) -> Any: """simple docstring""" a__ = True a__ = LlamaModel(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() a__ = model( __SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , encoder_hidden_states=__SCREAMING_SNAKE_CASE , encoder_attention_mask=__SCREAMING_SNAKE_CASE , ) a__ = model( __SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , encoder_hidden_states=__SCREAMING_SNAKE_CASE , ) a__ = model(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , ) -> List[str]: """simple docstring""" a__ = LlamaForCausalLM(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() a__ = model(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase__ ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , ) -> Optional[int]: """simple docstring""" a__ = True a__ = True a__ = LlamaForCausalLM(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() # first forward pass a__ = model( __SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , encoder_hidden_states=__SCREAMING_SNAKE_CASE , encoder_attention_mask=__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE , ) a__ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids a__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) a__ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and a__ = torch.cat([input_ids, next_tokens] , dim=-1 ) a__ = torch.cat([input_mask, next_mask] , dim=-1 ) a__ = model( __SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , encoder_hidden_states=__SCREAMING_SNAKE_CASE , encoder_attention_mask=__SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE , )['hidden_states'][0] a__ = model( __SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , encoder_hidden_states=__SCREAMING_SNAKE_CASE , encoder_attention_mask=__SCREAMING_SNAKE_CASE , past_key_values=__SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE , )['hidden_states'][0] # select random slice a__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() a__ = output_from_no_past[:, -3:, random_slice_idx].detach() a__ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1e-3 ) ) def lowercase__ ( self ) -> Dict: """simple docstring""" a__ = self.prepare_config_and_inputs() ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ) = config_and_inputs a__ = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class lowercase(_lowercase , _lowercase , _lowercase , unittest.TestCase ): __snake_case: Optional[int] = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () __snake_case: Optional[int] = (LlamaForCausalLM,) if is_torch_available() else () __snake_case: List[Any] = ( { 'feature-extraction': LlamaModel, 'text-classification': LlamaForSequenceClassification, 'text-generation': LlamaForCausalLM, 'zero-shot': LlamaForSequenceClassification, } if is_torch_available() else {} ) __snake_case: List[Any] = False __snake_case: List[str] = False def lowercase__ ( self ) -> Any: """simple docstring""" a__ = LlamaModelTester(self ) a__ = ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , hidden_size=3_7 ) def lowercase__ ( self ) -> List[Any]: """simple docstring""" self.config_tester.run_common_tests() def lowercase__ ( self ) -> Optional[Any]: """simple docstring""" a__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE ) def lowercase__ ( self ) -> Any: """simple docstring""" a__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: a__ = type self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE ) def lowercase__ ( self ) -> int: """simple docstring""" a__ , a__ = self.model_tester.prepare_config_and_inputs_for_common() a__ = 3 a__ = input_dict['input_ids'] a__ = input_ids.ne(1 ).to(__SCREAMING_SNAKE_CASE ) a__ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) a__ = LlamaForSequenceClassification(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() a__ = model(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowercase__ ( self ) -> Union[str, Any]: """simple docstring""" a__ , a__ = self.model_tester.prepare_config_and_inputs_for_common() a__ = 3 a__ = 'single_label_classification' a__ = input_dict['input_ids'] a__ = input_ids.ne(1 ).to(__SCREAMING_SNAKE_CASE ) a__ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) a__ = LlamaForSequenceClassification(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() a__ = model(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowercase__ ( self ) -> int: """simple docstring""" a__ , a__ = self.model_tester.prepare_config_and_inputs_for_common() a__ = 3 a__ = 'multi_label_classification' a__ = input_dict['input_ids'] a__ = input_ids.ne(1 ).to(__SCREAMING_SNAKE_CASE ) a__ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) a__ = LlamaForSequenceClassification(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() a__ = model(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def lowercase__ ( self ) -> List[str]: """simple docstring""" pass @parameterized.expand([('linear',), ('dynamic',)] ) def lowercase__ ( self , __SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" a__ , a__ = self.model_tester.prepare_config_and_inputs_for_common() a__ = ids_tensor([1, 1_0] , config.vocab_size ) a__ = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights a__ = LlamaModel(__SCREAMING_SNAKE_CASE ) original_model.to(__SCREAMING_SNAKE_CASE ) original_model.eval() a__ = original_model(__SCREAMING_SNAKE_CASE ).last_hidden_state a__ = original_model(__SCREAMING_SNAKE_CASE ).last_hidden_state set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights a__ = {'type': scaling_type, 'factor': 10.0} a__ = LlamaModel(__SCREAMING_SNAKE_CASE ) scaled_model.to(__SCREAMING_SNAKE_CASE ) scaled_model.eval() a__ = scaled_model(__SCREAMING_SNAKE_CASE ).last_hidden_state a__ = scaled_model(__SCREAMING_SNAKE_CASE ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1e-5 ) ) @require_torch class lowercase(unittest.TestCase ): @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowercase__ ( self ) -> Any: """simple docstring""" a__ = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] a__ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) a__ = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 a__ = torch.tensor([[-6.65_50, -4.12_27, -4.98_59, -3.24_06, 0.82_62, -3.00_33, 1.29_64, -3.36_99]] ) torch.testing.assert_close(out.mean(-1 ) , __SCREAMING_SNAKE_CASE , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off a__ = torch.tensor([-12.82_81, -7.44_53, -0.46_39, -8.06_25, -7.25_00, -8.00_00, -6.48_83, -7.76_95, -7.84_38, -7.03_12, -6.21_88, -7.13_28, -1.84_96, 1.99_61, -8.62_50, -6.72_27, -12.82_81, -6.94_92, -7.07_42, -7.78_52, -7.58_20, -7.90_62, -6.93_75, -7.98_05, -8.34_38, -8.15_62, -8.04_69, -7.62_50, -7.74_22, -7.33_98,] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , __SCREAMING_SNAKE_CASE , atol=1e-5 , rtol=1e-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowercase__ ( self ) -> Optional[int]: """simple docstring""" a__ = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] a__ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) a__ = model(torch.tensor(__SCREAMING_SNAKE_CASE ) ) # Expected mean on dim = -1 a__ = torch.tensor([[-2.06_22, -1.27_94, -1.16_38, -0.97_88, -1.46_03, -1.02_38, -1.78_93, -1.44_11]] ) torch.testing.assert_close(out.mean(-1 ) , __SCREAMING_SNAKE_CASE , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off a__ = torch.tensor([-8.14_06, -8.05_47, 2.74_61, -1.23_44, -0.14_48, -1.82_62, -1.00_20, -1.81_54, -1.68_95, -1.85_16, -2.35_74, -0.92_77, 3.75_98, 6.57_42, -1.29_98, -0.11_77, -8.14_06, -2.96_88, -2.91_99, -3.16_99, -3.52_54, -2.35_55, -2.79_88, -3.41_41, -2.82_62, -4.51_95, -3.33_79, -3.31_64, -2.78_32, -3.02_73] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , __SCREAMING_SNAKE_CASE , atol=1e-5 , rtol=1e-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def lowercase__ ( self ) -> str: """simple docstring""" a__ = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] a__ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) a__ = model(torch.tensor(__SCREAMING_SNAKE_CASE ) ) # Expected mean on dim = -1 a__ = torch.tensor([[-0.85_62, -1.85_20, -0.75_51, -0.41_62, -1.51_61, -1.20_38, -2.48_23, -2.32_54]] ) torch.testing.assert_close(out.mean(-1 ) , __SCREAMING_SNAKE_CASE , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off a__ = torch.tensor([-2.22_27, 4.88_28, 0.90_23, -0.45_78, -0.78_71, -0.10_33, -0.62_21, -0.57_86, -0.78_03, -1.06_74, -1.29_20, -0.15_70, 0.80_08, 2.07_23, -0.94_97, 0.27_71, -2.22_27, -0.76_12, -1.43_46, -1.20_61, -1.64_26, -0.30_00, -0.71_39, -1.19_34, -1.86_91, -1.69_73, -1.59_47, -1.27_05, -0.35_23, -0.55_13] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , __SCREAMING_SNAKE_CASE , atol=1e-2 , rtol=1e-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def lowercase__ ( self ) -> Tuple: """simple docstring""" a__ = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] a__ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) a__ = model(torch.tensor(__SCREAMING_SNAKE_CASE ) ) a__ = torch.tensor( [[-4.23_27, -3.33_60, -4.66_65, -4.76_31, -1.81_80, -3.41_70, -1.42_11, -3.18_10]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , __SCREAMING_SNAKE_CASE , atol=1e-2 , rtol=1e-2 ) # fmt: off a__ = torch.tensor([-9.49_22, -3.95_51, 1.79_98, -5.67_58, -5.10_55, -5.89_84, -4.83_20, -6.80_86, -6.53_91, -5.61_72, -5.58_20, -5.53_52, 1.78_81, 3.62_89, -6.51_17, -3.47_85, -9.50_00, -6.03_52, -6.81_25, -6.01_95, -6.68_36, -5.47_27, -6.28_12, -6.03_91, -7.33_98, -7.42_97, -7.48_44, -6.58_20, -5.87_89, -5.53_12] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , __SCREAMING_SNAKE_CASE , atol=1e-5 , rtol=1e-5 ) @unittest.skip('Model is curently gated' ) @slow def lowercase__ ( self ) -> List[str]: """simple docstring""" a__ = 'Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' a__ = 'Simply put, the theory of relativity states that ' a__ = LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) a__ = tokenizer.encode(__SCREAMING_SNAKE_CASE , return_tensors='pt' ) a__ = LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=__SCREAMING_SNAKE_CASE ) # greedy generation outputs a__ = model.generate(__SCREAMING_SNAKE_CASE , max_new_tokens=6_4 , top_p=__SCREAMING_SNAKE_CASE , temperature=1 , do_sample=__SCREAMING_SNAKE_CASE ) a__ = tokenizer.decode(generated_ids[0] , skip_special_tokens=__SCREAMING_SNAKE_CASE ) self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )
273
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __snake_case = { "configuration_llama": ["LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP", "LlamaConfig"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["LlamaTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["LlamaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ "LlamaForCausalLM", "LlamaModel", "LlamaPreTrainedModel", "LlamaForSequenceClassification", ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys __snake_case = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
181
import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class UpperCAmelCase : def __init__( self : Union[str, Any] , __magic_name__ : List[Any] , __magic_name__ : str=1_3 , __magic_name__ : str=7 , __magic_name__ : Optional[Any]=True , __magic_name__ : Optional[Any]=True , __magic_name__ : Tuple=9_9 , __magic_name__ : List[str]=3_2 , __magic_name__ : List[str]=5 , __magic_name__ : int=4 , __magic_name__ : Union[str, Any]=3_7 , __magic_name__ : List[Any]="gelu" , __magic_name__ : Any=0.1 , __magic_name__ : Union[str, Any]=0.1 , __magic_name__ : List[Any]=5_0 , __magic_name__ : Tuple=0.02 , __magic_name__ : Optional[int]=True , __magic_name__ : Optional[Any]=None , ): """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = seq_length UpperCamelCase = is_training UpperCamelCase = use_input_mask UpperCamelCase = vocab_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = initializer_range UpperCamelCase = use_labels UpperCamelCase = scope def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase = None if self.use_input_mask: UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase = self.get_config() return config, input_ids, input_mask, token_labels def lowerCamelCase_ ( self : str ): """simple docstring""" return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=__magic_name__ , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) = self.prepare_config_and_inputs() UpperCamelCase = True UpperCamelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def lowerCamelCase_ ( self : Tuple , __magic_name__ : List[Any] , __magic_name__ : Tuple , __magic_name__ : str , __magic_name__ : List[str] , **__magic_name__ : Optional[Any] , ): """simple docstring""" UpperCamelCase = BertGenerationEncoder(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() UpperCamelCase = model(__magic_name__ , attention_mask=__magic_name__ ) UpperCamelCase = model(__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self : Optional[int] , __magic_name__ : Dict , __magic_name__ : str , __magic_name__ : Union[str, Any] , __magic_name__ : List[str] , __magic_name__ : Optional[Any] , __magic_name__ : Tuple , **__magic_name__ : Union[str, Any] , ): """simple docstring""" UpperCamelCase = True UpperCamelCase = BertGenerationEncoder(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() UpperCamelCase = model( __magic_name__ , attention_mask=__magic_name__ , encoder_hidden_states=__magic_name__ , encoder_attention_mask=__magic_name__ , ) UpperCamelCase = model( __magic_name__ , attention_mask=__magic_name__ , encoder_hidden_states=__magic_name__ , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase_ ( self : Optional[int] , __magic_name__ : Optional[Any] , __magic_name__ : str , __magic_name__ : List[str] , __magic_name__ : Optional[int] , __magic_name__ : str , __magic_name__ : Union[str, Any] , **__magic_name__ : Dict , ): """simple docstring""" UpperCamelCase = True UpperCamelCase = True UpperCamelCase = BertGenerationDecoder(config=__magic_name__ ).to(__magic_name__ ).eval() # first forward pass UpperCamelCase = model( __magic_name__ , attention_mask=__magic_name__ , encoder_hidden_states=__magic_name__ , encoder_attention_mask=__magic_name__ , use_cache=__magic_name__ , ) UpperCamelCase = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCamelCase = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and UpperCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCamelCase = torch.cat([input_mask, next_mask] , dim=-1 ) UpperCamelCase = model( __magic_name__ , attention_mask=__magic_name__ , encoder_hidden_states=__magic_name__ , encoder_attention_mask=__magic_name__ , output_hidden_states=__magic_name__ , )["""hidden_states"""][0] UpperCamelCase = model( __magic_name__ , attention_mask=__magic_name__ , encoder_hidden_states=__magic_name__ , encoder_attention_mask=__magic_name__ , past_key_values=__magic_name__ , output_hidden_states=__magic_name__ , )["""hidden_states"""][0] # select random slice UpperCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCamelCase = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCamelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__magic_name__ , __magic_name__ , atol=1e-3 ) ) def lowerCamelCase_ ( self : int , __magic_name__ : Any , __magic_name__ : List[str] , __magic_name__ : Union[str, Any] , __magic_name__ : List[Any] , *__magic_name__ : Optional[Any] , ): """simple docstring""" UpperCamelCase = BertGenerationDecoder(__magic_name__ ) model.to(__magic_name__ ) model.eval() UpperCamelCase = model(__magic_name__ , attention_mask=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCAmelCase ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): lowercase = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () lowercase = (BertGenerationDecoder,) if is_torch_available() else () lowercase = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = BertGenerationEncoderTester(self ) UpperCamelCase = ConfigTester(self , config_class=__magic_name__ , hidden_size=3_7 ) def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" self.config_tester.run_common_tests() def lowerCamelCase_ ( self : Any ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs() UpperCamelCase = """bert""" self.model_tester.create_and_check_model(__magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__magic_name__ ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__magic_name__ ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() UpperCamelCase = None self.model_tester.create_and_check_model_as_decoder( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , ) def lowerCamelCase_ ( self : str ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*__magic_name__ ) @slow def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = BertGenerationEncoder.from_pretrained("""google/bert_for_seq_generation_L-24_bbc_encoder""" ) self.assertIsNotNone(__magic_name__ ) @require_torch class UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = BertGenerationEncoder.from_pretrained("""google/bert_for_seq_generation_L-24_bbc_encoder""" ) UpperCamelCase = torch.tensor([[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_2_6, 3_8_9_9, 2_0_0_3, 1_0_1_4_0, 1_0_2]] ) with torch.no_grad(): UpperCamelCase = model(__magic_name__ )[0] UpperCamelCase = torch.Size([1, 8, 1_0_2_4] ) self.assertEqual(output.shape , __magic_name__ ) UpperCamelCase = torch.tensor( [[[0.1_775, 0.0_083, -0.0_321], [1.6_002, 0.1_287, 0.3_912], [2.1_473, 0.5_791, 0.6_066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __magic_name__ , atol=1e-4 ) ) @require_torch class UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = BertGenerationDecoder.from_pretrained("""google/bert_for_seq_generation_L-24_bbc_encoder""" ) UpperCamelCase = torch.tensor([[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_2_6, 3_8_9_9, 2_0_0_3, 1_0_1_4_0, 1_0_2]] ) with torch.no_grad(): UpperCamelCase = model(__magic_name__ )[0] UpperCamelCase = torch.Size([1, 8, 5_0_3_5_8] ) self.assertEqual(output.shape , __magic_name__ ) UpperCamelCase = torch.tensor( [[[-0.5_788, -2.5_994, -3.7_054], [0.0_438, 4.7_997, 1.8_795], [1.5_862, 6.6_409, 4.4_638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __magic_name__ , atol=1e-4 ) )
181
1
import itertools import os import random import tempfile import unittest import numpy as np from transformers import TvltFeatureExtractor, is_datasets_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch if is_datasets_available(): from datasets import load_dataset a_ = random.Random() def lowerCamelCase__ ( _a , _a=1.0 , _a=None , _a=None): if rng is None: SCREAMING_SNAKE_CASE : List[str] = global_rng SCREAMING_SNAKE_CASE : Optional[int] = [] for batch_idx in range(shape[0]): values.append([]) for _ in range(shape[1]): values[-1].append(rng.random() * scale) return values class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : str , a : Any , a : Union[str, Any]=7 , a : List[Any]=400 , a : str=2000 , a : Dict=2048 , a : List[Any]=128 , a : Tuple=1 , a : Union[str, Any]=512 , a : List[str]=30 , a : Tuple=4_4100 , ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = parent SCREAMING_SNAKE_CASE : List[str] = batch_size SCREAMING_SNAKE_CASE : List[str] = min_seq_length SCREAMING_SNAKE_CASE : List[str] = max_seq_length SCREAMING_SNAKE_CASE : Optional[Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) SCREAMING_SNAKE_CASE : Dict = spectrogram_length SCREAMING_SNAKE_CASE : Optional[int] = feature_size SCREAMING_SNAKE_CASE : List[Any] = num_audio_channels SCREAMING_SNAKE_CASE : Optional[Any] = hop_length SCREAMING_SNAKE_CASE : List[Any] = chunk_length SCREAMING_SNAKE_CASE : List[str] = sampling_rate def __UpperCamelCase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" return { "spectrogram_length": self.spectrogram_length, "feature_size": self.feature_size, "num_audio_channels": self.num_audio_channels, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "sampling_rate": self.sampling_rate, } def __UpperCamelCase ( self : Optional[int] , a : int=False , a : Tuple=False ) -> Union[str, Any]: """simple docstring""" def _flatten(a : Any ): return list(itertools.chain(*a ) ) if equal_length: SCREAMING_SNAKE_CASE : Optional[int] = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size SCREAMING_SNAKE_CASE : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: SCREAMING_SNAKE_CASE : Any = [np.asarray(a ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _UpperCamelCase ( __A , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =TvltFeatureExtractor def __UpperCamelCase ( self : int ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE : int = TvltFeatureExtractionTester(self ) def __UpperCamelCase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(a , "spectrogram_length" ) ) self.assertTrue(hasattr(a , "feature_size" ) ) self.assertTrue(hasattr(a , "num_audio_channels" ) ) self.assertTrue(hasattr(a , "hop_length" ) ) self.assertTrue(hasattr(a , "chunk_length" ) ) self.assertTrue(hasattr(a , "sampling_rate" ) ) def __UpperCamelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE : str = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE : List[Any] = feat_extract_first.save_pretrained(a )[0] check_json_file_has_correct_format(a ) SCREAMING_SNAKE_CASE : Any = self.feature_extraction_class.from_pretrained(a ) SCREAMING_SNAKE_CASE : Optional[Any] = feat_extract_first.to_dict() SCREAMING_SNAKE_CASE : int = feat_extract_second.to_dict() SCREAMING_SNAKE_CASE : str = dict_first.pop("mel_filters" ) SCREAMING_SNAKE_CASE : str = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(a , a ) ) self.assertEqual(a , a ) def __UpperCamelCase ( self : Any ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join(a , "feat_extract.json" ) feat_extract_first.to_json_file(a ) SCREAMING_SNAKE_CASE : Optional[Any] = self.feature_extraction_class.from_json_file(a ) SCREAMING_SNAKE_CASE : Optional[Any] = feat_extract_first.to_dict() SCREAMING_SNAKE_CASE : Tuple = feat_extract_second.to_dict() SCREAMING_SNAKE_CASE : str = dict_first.pop("mel_filters" ) SCREAMING_SNAKE_CASE : Optional[Any] = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(a , a ) ) self.assertEqual(a , a ) def __UpperCamelCase ( self : Dict ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : int = self.feature_extraction_class(**self.feat_extract_dict ) # create three inputs of length 800, 1000, and 1200 SCREAMING_SNAKE_CASE : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] SCREAMING_SNAKE_CASE : str = [np.asarray(a ) for speech_input in speech_inputs] # Test not batched input SCREAMING_SNAKE_CASE : str = feature_extractor(np_speech_inputs[0] , return_tensors="np" , sampling_rate=4_4100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test batched SCREAMING_SNAKE_CASE : Tuple = feature_extractor(a , return_tensors="np" , sampling_rate=4_4100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test audio masking SCREAMING_SNAKE_CASE : Union[str, Any] = feature_extractor( a , return_tensors="np" , sampling_rate=4_4100 , mask_audio=a ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test 2-D numpy arrays are batched. SCREAMING_SNAKE_CASE : Any = [floats_list((1, x) )[0] for x in (800, 800, 800)] SCREAMING_SNAKE_CASE : Optional[Any] = np.asarray(a ) SCREAMING_SNAKE_CASE : Optional[int] = feature_extractor(a , return_tensors="np" , sampling_rate=4_4100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) def __UpperCamelCase ( self : List[Any] , a : str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech SCREAMING_SNAKE_CASE : Dict = ds.sort("id" ).select(range(a ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def __UpperCamelCase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self._load_datasamples(1 ) SCREAMING_SNAKE_CASE : Tuple = TvltFeatureExtractor() SCREAMING_SNAKE_CASE : Any = feature_extractor(a , return_tensors="pt" ).audio_values self.assertEquals(audio_values.shape , (1, 1, 192, 128) ) SCREAMING_SNAKE_CASE : List[str] = torch.tensor([[-0.3032, -0.2708], [-0.4434, -0.4007]] ) self.assertTrue(torch.allclose(audio_values[0, 0, :2, :2] , a , atol=1e-4 ) )
25
from __future__ import annotations import numpy as np def __a ( __lowerCAmelCase ) -> Optional[Any]: return np.maximum(0 , __lowerCAmelCase ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
352
0
'''simple docstring''' import math _lowerCAmelCase = 10 _lowerCAmelCase = 7 _lowerCAmelCase = BALLS_PER_COLOUR * NUM_COLOURS def __lowerCAmelCase ( snake_case__ = 20 ): __UpperCamelCase : Optional[Any] = math.comb(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCamelCase : int = math.comb(NUM_BALLS - BALLS_PER_COLOUR , __SCREAMING_SNAKE_CASE ) __UpperCamelCase : Optional[Any] = NUM_COLOURS * (1 - missing_colour / total) return F"{result:.9f}" if __name__ == "__main__": print(solution(20))
704
'''simple docstring''' def __lowerCAmelCase ( snake_case__ ): return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(snake_case__ ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__('''doctest''').testmod()
399
0
'''simple docstring''' def A__ ( ): for n in range(1 , 100_0000 ): yield n * (n + 1) // 2 def A__ ( __lowerCAmelCase : Union[str, Any] ): lowerCamelCase__ = 1 lowerCamelCase__ = 2 while i * i <= n: lowerCamelCase__ = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def A__ ( ): return next(i for i in triangle_number_generator() if count_divisors(__lowerCAmelCase ) > 500 ) if __name__ == "__main__": print(solution())
50
'''simple docstring''' def lowerCamelCase_ ( __UpperCamelCase : int , __UpperCamelCase : int ) -> int: """simple docstring""" while second != 0: _A = first & second first ^= second _A = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase = int(input("""Enter the first number: """).strip()) lowerCAmelCase = int(input("""Enter the second number: """).strip()) print(F'{add(first, second) = }')
292
0
"""simple docstring""" import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union _lowerCAmelCase = re.compile(r"^(?P<major>\d+)" r"\.(?P<minor>\d+)" r"\.(?P<patch>\d+)$") @total_ordering @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" _A : str _A : Optional[str] = None _A : Optional[Union[str, int]] = None _A : Optional[Union[str, int]] = None _A : Optional[Union[str, int]] = None def lowerCamelCase(self ): A_ : Union[str, Any] = _str_to_version_tuple(self.version_str ) def __repr__(self ): return f"""{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}""" @property def lowerCamelCase(self ): return self.major, self.minor, self.patch def lowerCamelCase(self , lowerCAmelCase_ ): if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return Version(lowerCAmelCase_ ) elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return other raise TypeError(f"""{other} (type {type(lowerCAmelCase_ )}) cannot be compared to version.""" ) def __eq__(self , lowerCAmelCase_ ): try: A_ : Optional[Any] = self._validate_operand(lowerCAmelCase_ ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__(self , lowerCAmelCase_ ): A_ : Tuple = self._validate_operand(lowerCAmelCase_ ) return self.tuple < other.tuple def __hash__(self ): return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def lowerCamelCase(cls , lowerCAmelCase_ ): A_ : Union[str, Any] = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def lowerCamelCase(self ): return self.version_str def __UpperCamelCase ( snake_case__ ): A_ : Any = _VERSION_REG.match(snake_case__ ) if not res: raise ValueError(F"""Invalid version '{version_str}'. Format should be x.y.z with {{x,y,z}} being digits.""" ) return tuple(int(snake_case__ ) for v in [res.group("""major""" ), res.group("""minor""" ), res.group("""patch""" )] ) def __UpperCamelCase ( snake_case__ ): return ".".join(str(snake_case__ ) for v in version_tuple )
713
"""simple docstring""" import unittest import torch from torch import nn from diffusers.models.activations import get_activation class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def lowerCamelCase(self ): A_ : Optional[int] = get_activation("""swish""" ) self.assertIsInstance(lowerCAmelCase_ , nn.SiLU ) self.assertEqual(act(torch.tensor(-100 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def lowerCamelCase(self ): A_ : Dict = get_activation("""silu""" ) self.assertIsInstance(lowerCAmelCase_ , nn.SiLU ) self.assertEqual(act(torch.tensor(-100 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def lowerCamelCase(self ): A_ : Optional[Any] = get_activation("""mish""" ) self.assertIsInstance(lowerCAmelCase_ , nn.Mish ) self.assertEqual(act(torch.tensor(-200 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 ) def lowerCamelCase(self ): A_ : int = get_activation("""gelu""" ) self.assertIsInstance(lowerCAmelCase_ , nn.GELU ) self.assertEqual(act(torch.tensor(-100 , dtype=torch.floataa ) ).item() , 0 ) self.assertNotEqual(act(torch.tensor(-1 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(0 , dtype=torch.floataa ) ).item() , 0 ) self.assertEqual(act(torch.tensor(20 , dtype=torch.floataa ) ).item() , 20 )
480
0
def a__ ( lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ): raise TypeError("Input value must be an 'int' type" ) UpperCAmelCase_ =0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
54
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor __lowercase : List[Any] =logging.get_logger(__name__) class A ( __lowercase ): def __init__( self: List[Any] , *_lowerCAmelCase: Optional[Any] , **_lowerCAmelCase: List[str] ) -> None: '''simple docstring''' warnings.warn( "The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use GLPNImageProcessor instead." , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
54
1
'''simple docstring''' # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __snake_case = abspath(join(dirname(__file__), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def a ( __a ) -> Optional[Any]: '''simple docstring''' config.addinivalue_line( '''markers''' , '''is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested''' ) config.addinivalue_line( '''markers''' , '''is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested''' ) config.addinivalue_line('''markers''' , '''is_pipeline_test: mark test to run only when pipelines are tested''' ) config.addinivalue_line('''markers''' , '''is_staging_test: mark test to run only in the staging environment''' ) config.addinivalue_line('''markers''' , '''accelerate_tests: mark test that require accelerate''' ) config.addinivalue_line('''markers''' , '''tool_tests: mark the tool tests that are run on their specific schedule''' ) def a ( __a ) -> Union[str, Any]: '''simple docstring''' from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(_snake_case ) def a ( __a ) -> Union[str, Any]: '''simple docstring''' from transformers.testing_utils import pytest_terminal_summary_main UpperCamelCase__ :Dict = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(_snake_case , id=_snake_case ) def a ( __a , __a ) -> int: '''simple docstring''' if exitstatus == 5: UpperCamelCase__ :Dict = 0 # Doctest custom flag to ignore output. __snake_case = doctest.register_optionflag('''IGNORE_RESULT''') __snake_case = doctest.OutputChecker class lowercase ( __UpperCAmelCase ): """simple docstring""" def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) __snake_case = CustomOutputChecker __snake_case = HfDoctestModule __snake_case = HfDocTestParser
711
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging __snake_case = logging.get_logger(__name__) if is_vision_available(): import PIL class lowercase ( A__ ): """simple docstring""" _a = ['pixel_values'] def __init__( self , UpperCamelCase_ = True , UpperCamelCase_ = None , UpperCamelCase_ = PILImageResampling.BICUBIC , UpperCamelCase_ = True , UpperCamelCase_ = None , UpperCamelCase_ = True , UpperCamelCase_ = 1 / 255 , UpperCamelCase_ = True , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = True , **UpperCamelCase_ , ): '''simple docstring''' super().__init__(**UpperCamelCase_ ) UpperCamelCase__ :Tuple = size if size is not None else {'''shortest_edge''': 224} UpperCamelCase__ :Optional[int] = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) UpperCamelCase__ :str = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} UpperCamelCase__ :Union[str, Any] = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ , param_name='''crop_size''' ) UpperCamelCase__ :Any = do_resize UpperCamelCase__ :Union[str, Any] = size UpperCamelCase__ :Any = resample UpperCamelCase__ :Optional[Any] = do_center_crop UpperCamelCase__ :List[str] = crop_size UpperCamelCase__ :Optional[int] = do_rescale UpperCamelCase__ :Optional[Any] = rescale_factor UpperCamelCase__ :Any = do_normalize UpperCamelCase__ :int = image_mean if image_mean is not None else OPENAI_CLIP_MEAN UpperCamelCase__ :List[str] = image_std if image_std is not None else OPENAI_CLIP_STD UpperCamelCase__ :Union[str, Any] = do_convert_rgb def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = PILImageResampling.BICUBIC , UpperCamelCase_ = None , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) if "shortest_edge" not in size: raise ValueError(F'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) UpperCamelCase__ :str = get_resize_output_image_size(UpperCamelCase_ , size=size['''shortest_edge'''] , default_to_square=UpperCamelCase_ ) return resize(UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :int = get_size_dict(UpperCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(F'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(UpperCamelCase_ , size=(size['''height'''], size['''width''']) , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None , **UpperCamelCase_ , ): '''simple docstring''' return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None , **UpperCamelCase_ , ): '''simple docstring''' return normalize(UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = ChannelDimension.FIRST , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :Optional[int] = do_resize if do_resize is not None else self.do_resize UpperCamelCase__ :Optional[Any] = size if size is not None else self.size UpperCamelCase__ :Optional[int] = get_size_dict(UpperCamelCase_ , param_name='''size''' , default_to_square=UpperCamelCase_ ) UpperCamelCase__ :Dict = resample if resample is not None else self.resample UpperCamelCase__ :int = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase__ :Any = crop_size if crop_size is not None else self.crop_size UpperCamelCase__ :Any = get_size_dict(UpperCamelCase_ , param_name='''crop_size''' , default_to_square=UpperCamelCase_ ) UpperCamelCase__ :List[str] = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase__ :List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase__ :Optional[int] = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase__ :Tuple = image_mean if image_mean is not None else self.image_mean UpperCamelCase__ :str = image_std if image_std is not None else self.image_std UpperCamelCase__ :Optional[Any] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb UpperCamelCase__ :str = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: UpperCamelCase__ :Any = [convert_to_rgb(UpperCamelCase_ ) for image in images] # All transformations expect numpy arrays. UpperCamelCase__ :str = [to_numpy_array(UpperCamelCase_ ) for image in images] if do_resize: UpperCamelCase__ :Optional[Any] = [self.resize(image=UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ ) for image in images] if do_center_crop: UpperCamelCase__ :Dict = [self.center_crop(image=UpperCamelCase_ , size=UpperCamelCase_ ) for image in images] if do_rescale: UpperCamelCase__ :Optional[Any] = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_normalize: UpperCamelCase__ :Tuple = [self.normalize(image=UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ ) for image in images] UpperCamelCase__ :List[str] = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] UpperCamelCase__ :Optional[Any] = {'''pixel_values''': images} return BatchFeature(data=UpperCamelCase_ , tensor_type=UpperCamelCase_ )
280
0
import os import sys __a = os.path.join(os.path.dirname(__file__), '''src''') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) __a = [ '''torch''', '''numpy''', '''tokenizers''', '''filelock''', '''requests''', '''tqdm''', '''regex''', '''sentencepiece''', '''sacremoses''', '''importlib_metadata''', '''huggingface_hub''', ] @add_start_docstrings(AutoConfig.__doc__ ) def __lowercase ( *_UpperCamelCase, **_UpperCamelCase ) ->Dict: """simple docstring""" return AutoConfig.from_pretrained(*_UpperCamelCase, **_UpperCamelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def __lowercase ( *_UpperCamelCase, **_UpperCamelCase ) ->Union[str, Any]: """simple docstring""" return AutoTokenizer.from_pretrained(*_UpperCamelCase, **_UpperCamelCase ) @add_start_docstrings(AutoModel.__doc__ ) def __lowercase ( *_UpperCamelCase, **_UpperCamelCase ) ->Union[str, Any]: """simple docstring""" return AutoModel.from_pretrained(*_UpperCamelCase, **_UpperCamelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def __lowercase ( *_UpperCamelCase, **_UpperCamelCase ) ->str: """simple docstring""" return AutoModelForCausalLM.from_pretrained(*_UpperCamelCase, **_UpperCamelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def __lowercase ( *_UpperCamelCase, **_UpperCamelCase ) ->Tuple: """simple docstring""" return AutoModelForMaskedLM.from_pretrained(*_UpperCamelCase, **_UpperCamelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def __lowercase ( *_UpperCamelCase, **_UpperCamelCase ) ->Optional[Any]: """simple docstring""" return AutoModelForSequenceClassification.from_pretrained(*_UpperCamelCase, **_UpperCamelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def __lowercase ( *_UpperCamelCase, **_UpperCamelCase ) ->List[Any]: """simple docstring""" return AutoModelForQuestionAnswering.from_pretrained(*_UpperCamelCase, **_UpperCamelCase )
319
import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @require_torch def __lowerCamelCase ( self ): lowercase : Optional[Any] = pipeline( task='''zero-shot-audio-classification''' , model='''hf-internal-testing/tiny-clap-htsat-unfused''' ) lowercase : Any = load_dataset('''ashraq/esc50''' ) lowercase : Union[str, Any] = dataset['''train''']['''audio'''][-1]['''array'''] lowercase : Optional[int] = audio_classifier(SCREAMING_SNAKE_CASE__ , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [{'''score''': 0.501, '''label''': '''Sound of a dog'''}, {'''score''': 0.499, '''label''': '''Sound of vaccum cleaner'''}] , ) @unittest.skip('''No models are available in TF''' ) def __lowerCamelCase ( self ): pass @slow @require_torch def __lowerCamelCase ( self ): lowercase : List[str] = pipeline( task='''zero-shot-audio-classification''' , model='''laion/clap-htsat-unfused''' , ) # This is an audio of a dog lowercase : List[str] = load_dataset('''ashraq/esc50''' ) lowercase : Dict = dataset['''train''']['''audio'''][-1]['''array'''] lowercase : Dict = audio_classifier(SCREAMING_SNAKE_CASE__ , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [ {'''score''': 0.999, '''label''': '''Sound of a dog'''}, {'''score''': 0.001, '''label''': '''Sound of vaccum cleaner'''}, ] , ) lowercase : Tuple = audio_classifier([audio] * 5 , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [ [ {'''score''': 0.999, '''label''': '''Sound of a dog'''}, {'''score''': 0.001, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 , ) lowercase : Dict = audio_classifier( [audio] * 5 , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] , batch_size=5 ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE__ ) , [ [ {'''score''': 0.999, '''label''': '''Sound of a dog'''}, {'''score''': 0.001, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 , ) @unittest.skip('''No models are available in TF''' ) def __lowerCamelCase ( self ): pass
319
1
'''simple docstring''' from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def __UpperCAmelCase ( A : float , A : float , A : bool = False ) -> list[float]: if radian_mode: return [magnitude * cos(A ), magnitude * sin(A )] return [magnitude * cos(radians(A ) ), magnitude * sin(radians(A ) )] def __UpperCAmelCase ( A : NDArray[floataa] , A : NDArray[floataa] , A : float = 1_0**-1 ) -> bool: UpperCAmelCase_ : NDArray[floataa] = cross(A , A ) UpperCAmelCase_ : float = sum(A ) return abs(A ) < eps if __name__ == "__main__": # Test to check if it works _UpperCamelCase : Optional[int] = array( [ polar_force(7_18.4, 180 - 30), polar_force(8_79.54, 45), polar_force(100, -90), ] ) _UpperCamelCase : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg _UpperCamelCase : List[Any] = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) _UpperCamelCase : Union[str, Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg _UpperCamelCase : int = array([[0, -2_000], [0, -1_200], [0, 15_600], [0, -12_400]]) _UpperCamelCase : Union[str, Any] = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
216
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCamelCase : Optional[int] = { 'configuration_longt5': ['LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LongT5Config', 'LongT5OnnxConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : List[str] = [ 'LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'LongT5EncoderModel', 'LongT5ForConditionalGeneration', 'LongT5Model', 'LongT5PreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : int = [ 'FlaxLongT5ForConditionalGeneration', 'FlaxLongT5Model', 'FlaxLongT5PreTrainedModel', ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCamelCase : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
216
1
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE__ ( metaclass=UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = ['''speech'''] def __init__( self,*__lowerCamelCase,**__lowerCamelCase ): requires_backends(self,['''speech'''] ) class SCREAMING_SNAKE_CASE__ ( metaclass=UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = ['''speech'''] def __init__( self,*__lowerCamelCase,**__lowerCamelCase ): requires_backends(self,['''speech'''] )
190
import pytest import datasets # Import fixture modules as plugins a__: Dict = ['tests.fixtures.files', 'tests.fixtures.hub', 'tests.fixtures.fsspec'] def UpperCamelCase__( UpperCamelCase__ : List[Any] , UpperCamelCase__ : Tuple )->List[str]: # Mark tests as "unit" by default if not marked as "integration" (or already marked as "unit") for item in items: if any(marker in item.keywords for marker in ['''integration''', '''unit'''] ): continue item.add_marker(pytest.mark.unit ) def UpperCamelCase__( UpperCamelCase__ : Dict )->List[Any]: config.addinivalue_line('''markers''' , '''torchaudio_latest: mark test to run with torchaudio>=0.12''' ) @pytest.fixture(autouse=UpperCamelCase__ ) def UpperCamelCase__( UpperCamelCase__ : Optional[int] , UpperCamelCase__ : int )->Union[str, Any]: # test_hf_cache_home = tmp_path_factory.mktemp("cache") # TODO: why a cache dir per test function does not work? A__ = tmp_path_factory.getbasetemp() / '''cache''' A__ = test_hf_cache_home / '''datasets''' A__ = test_hf_cache_home / '''metrics''' A__ = test_hf_cache_home / '''modules''' monkeypatch.setattr('''datasets.config.HF_DATASETS_CACHE''' , str(UpperCamelCase__ ) ) monkeypatch.setattr('''datasets.config.HF_METRICS_CACHE''' , str(UpperCamelCase__ ) ) monkeypatch.setattr('''datasets.config.HF_MODULES_CACHE''' , str(UpperCamelCase__ ) ) A__ = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr('''datasets.config.DOWNLOADED_DATASETS_PATH''' , str(UpperCamelCase__ ) ) A__ = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(UpperCamelCase__ ) ) @pytest.fixture(autouse=UpperCamelCase__ , scope='''session''' ) def UpperCamelCase__( )->int: datasets.disable_progress_bar() @pytest.fixture(autouse=UpperCamelCase__ ) def UpperCamelCase__( UpperCamelCase__ : str )->Tuple: # don't take tests into account when counting downloads monkeypatch.setattr('''datasets.config.HF_UPDATE_DOWNLOAD_COUNTS''' , UpperCamelCase__ ) @pytest.fixture def UpperCamelCase__( UpperCamelCase__ : List[Any] )->Optional[Any]: # Required to suppress RemovedIn20Warning when feature(s) are not compatible with SQLAlchemy 2.0 # To be removed once SQLAlchemy 2.0 supported monkeypatch.setattr('''sqlalchemy.util.deprecations.SILENCE_UBER_WARNING''' , UpperCamelCase__ )
190
1
import unittest from pathlib import Path from shutil import copyfile from transformers import SPIECE_UNDERLINE, is_sentencepiece_available from transformers.models.speech_to_text import SpeechaTextTokenizer from transformers.models.speech_to_text.tokenization_speech_to_text import VOCAB_FILES_NAMES, save_json from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ = get_tests_dir('fixtures/test_sentencepiece.model') if is_sentencepiece_available(): import sentencepiece as sp lowerCAmelCase__ = 5 lowerCAmelCase__ = 10 @require_sentencepiece @require_tokenizers class snake_case ( lowercase_ ,unittest.TestCase ): """simple docstring""" __lowerCAmelCase = SpeechaTextTokenizer __lowerCAmelCase = False __lowerCAmelCase = True def snake_case__ ( self ): super().setUp() __lowercase = sp.SentencePieceProcessor() spm_model.Load(lowerCamelCase_ ) __lowercase = ["""<s>""", """<pad>""", """</s>""", """<unk>"""] vocab += [spm_model.IdToPiece(id_ ) for id_ in range(len(lowerCamelCase_ ) )] __lowercase = dict(zip(lowerCamelCase_ , range(len(lowerCamelCase_ ) ) ) ) __lowercase = Path(self.tmpdirname ) save_json(lowerCamelCase_ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(lowerCamelCase_ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) __lowercase = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self ): __lowercase = """<pad>""" __lowercase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase_ ) , lowerCamelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase_ ) , lowerCamelCase_ ) def snake_case__ ( self ): __lowercase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(lowerCamelCase_ ) , 1001 ) def snake_case__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1001 ) def snake_case__ ( self ): __lowercase = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) __lowercase = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCamelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , [289, 50, 14, 174, 386] , ) __lowercase = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCamelCase_ , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", "."] , ) __lowercase = tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , [12, 25, 88, 59, 28, 23, 11, 4, 606, 351, 351, 351, 7, 16, 70, 50, 76, 84, 10, 4, 8] ) __lowercase = tokenizer.convert_ids_to_tokens(lowerCamelCase_ ) self.assertListEqual( lowerCamelCase_ , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", "."] , ) @slow def snake_case__ ( self ): __lowercase = {"""input_ids""": [[3791, 797, 31, 11, 64, 797, 31, 2429, 433, 12, 1176, 12, 20, 786, 915, 142, 2413, 240, 37, 3238, 797, 31, 11, 35, 93, 915, 142, 2413, 240, 37, 5540, 567, 1276, 93, 37, 610, 40, 62, 455, 657, 1042, 123, 780, 177, 37, 309, 241, 1298, 514, 20, 292, 2737, 114, 2469, 241, 85, 64, 302, 548, 528, 423, 4, 509, 406, 423, 37, 601, 4, 777, 302, 548, 528, 423, 284, 4, 3388, 511, 459, 4, 3555, 40, 321, 302, 705, 4, 3388, 511, 583, 326, 5, 5, 5, 62, 3310, 560, 177, 2680, 217, 1508, 32, 31, 853, 418, 64, 583, 511, 1605, 62, 35, 93, 560, 177, 2680, 217, 1508, 1521, 64, 583, 511, 519, 62, 20, 1515, 764, 20, 149, 261, 5625, 7972, 20, 5540, 567, 1276, 93, 3925, 1675, 11, 15, 802, 7972, 576, 217, 1508, 11, 35, 93, 1253, 2441, 15, 289, 652, 31, 416, 321, 3842, 115, 40, 911, 8, 476, 619, 4, 380, 142, 423, 335, 240, 35, 93, 264, 8, 11, 335, 569, 420, 163, 5, 2], [260, 548, 528, 423, 20, 451, 20, 2681, 1153, 3434, 20, 5540, 37, 567, 126, 1253, 2441, 3376, 449, 210, 431, 1563, 177, 767, 5540, 11, 1203, 472, 11, 2953, 685, 285, 364, 706, 1153, 20, 6799, 20, 2869, 20, 4464, 126, 40, 2429, 20, 1040, 866, 2664, 418, 20, 318, 20, 1726, 186, 20, 265, 522, 35, 93, 2191, 4634, 20, 1040, 12, 6799, 15, 228, 2356, 142, 31, 11, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [2575, 2666, 684, 1582, 1176, 12, 627, 149, 619, 20, 4902, 563, 11, 20, 149, 261, 3420, 2356, 174, 142, 4714, 131, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase_ , model_name="facebook/s2t-small-mustc-en-de-st" , revision="a14f04cf0776c02f62a8cb800cf7909e15ea23ad" , ) @require_sentencepiece class snake_case ( unittest.TestCase ): """simple docstring""" __lowerCAmelCase = """valhalla/s2t_mustc_multilinguial_medium""" __lowerCAmelCase = """C\'est trop cool""" __lowerCAmelCase = """Esto es genial""" @classmethod def snake_case__ ( cls ): __lowercase = SpeechaTextTokenizer.from_pretrained(cls.checkpoint_name ) return cls def snake_case__ ( self ): self.assertEqual(self.tokenizer.lang_code_to_id["pt"] , 4 ) self.assertEqual(self.tokenizer.lang_code_to_id["ru"] , 6 ) self.assertEqual(self.tokenizer.lang_code_to_id["it"] , 9 ) self.assertEqual(self.tokenizer.lang_code_to_id["de"] , 11 ) def snake_case__ ( self ): self.assertEqual(self.tokenizer.vocab_size , 1_0000 ) def snake_case__ ( self ): self.assertIn(lowerCamelCase_ , self.tokenizer.all_special_ids ) __lowercase = [ES_CODE, 4, 1601, 47, 7647, 2] __lowercase = self.tokenizer.decode(lowerCamelCase_ , skip_special_tokens=lowerCamelCase_ ) __lowercase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=lowerCamelCase_ ) self.assertEqual(lowerCamelCase_ , lowerCamelCase_ ) self.assertNotIn(self.tokenizer.eos_token , lowerCamelCase_ ) def snake_case__ ( self ): __lowercase = """fr""" __lowercase = self.tokenizer(self.french_text ).input_ids self.assertEqual(encoded[0] , lowerCamelCase_ ) self.assertEqual(encoded[-1] , self.tokenizer.eos_token_id ) def snake_case__ ( self ): __lowercase = """fr""" self.assertListEqual(self.tokenizer.prefix_tokens , [FR_CODE] ) __lowercase = """es""" self.assertListEqual(self.tokenizer.prefix_tokens , [ES_CODE] )
713
import math def __lowercase ( _UpperCAmelCase ) -> str: '''simple docstring''' __lowercase = 0 __lowercase = 0 while num > 0: __lowercase = num % 8 __lowercase = octal + (remainder * math.floor(math.pow(10 , _UpperCAmelCase ) )) counter += 1 __lowercase = math.floor(num / 8 ) # basically /= 8 without remainder if any # This formatting removes trailing '.0' from `octal`. return f'''0o{int(_UpperCAmelCase )}''' def __lowercase ( ) -> None: '''simple docstring''' print("\n2 in octal is:" ) print(decimal_to_octal(2 ) ) # = 2 print("\n8 in octal is:" ) print(decimal_to_octal(8 ) ) # = 10 print("\n65 in octal is:" ) print(decimal_to_octal(65 ) ) # = 101 print("\n216 in octal is:" ) print(decimal_to_octal(216 ) ) # = 330 print("\n512 in octal is:" ) print(decimal_to_octal(512 ) ) # = 1000 print("\n" ) if __name__ == "__main__": main()
576
0
def lowerCAmelCase__ ( lowerCamelCase_ : str ,lowerCamelCase_ : Tuple): '''simple docstring''' if density <= 0: raise ValueError('''Impossible fluid density''') if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''') return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
647
import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ = logging.get_logger(__name__) def UpperCAmelCase_ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE__ =torch.load(__UpperCamelCase, map_location="""cpu""" ) if "model" in sd.keys(): SCREAMING_SNAKE_CASE__ =torch.load(__UpperCamelCase, map_location="""cpu""" )["""model"""] # pop unnecessary weights SCREAMING_SNAKE_CASE__ =[ """decoder.version""", """decoder.output_projection.weight""", ] for key in keys_to_delete: if key in sd: sd.pop(__UpperCamelCase ) SCREAMING_SNAKE_CASE__ ={ """decoder.project_in_dim.weight""": """decoder.project_in.weight""", """decoder.project_out_dim.weight""": """decoder.project_out.weight""", """decoder.layer_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.layer_norm.bias""": """decoder.final_layer_norm.bias""", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: SCREAMING_SNAKE_CASE__ =sd.pop(__UpperCamelCase ) SCREAMING_SNAKE_CASE__ =list(sd.keys() ) for key in keys: if ".qkv_proj." in key: SCREAMING_SNAKE_CASE__ =sd[key] # We split QKV in separate Q,K,V SCREAMING_SNAKE_CASE__ =key.replace(""".qkv_proj.""", """.q_proj.""" ) SCREAMING_SNAKE_CASE__ =key.replace(""".qkv_proj.""", """.k_proj.""" ) SCREAMING_SNAKE_CASE__ =key.replace(""".qkv_proj.""", """.v_proj.""" ) SCREAMING_SNAKE_CASE__ =value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ =torch.split(__UpperCamelCase, depth // 3, dim=0 ) SCREAMING_SNAKE_CASE__ =q SCREAMING_SNAKE_CASE__ =k SCREAMING_SNAKE_CASE__ =v del sd[key] return sd @torch.no_grad() def UpperCAmelCase_ ( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase=None ): SCREAMING_SNAKE_CASE__ =load_checkpoint(__UpperCamelCase ) if config is not None: SCREAMING_SNAKE_CASE__ =OPTConfig.from_pretrained(__UpperCamelCase ) else: SCREAMING_SNAKE_CASE__ =OPTConfig() SCREAMING_SNAKE_CASE__ =OPTModel(__UpperCamelCase ).half().eval() model.load_state_dict(__UpperCamelCase ) # Check results Path(__UpperCamelCase ).mkdir(exist_ok=__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--fairseq_path", type=str, help=( "path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:" " https://huggingface.co/models?other=opt_metasq" ), ) parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--hf_config", default=None, type=str, help="Define HF config.") lowerCamelCase_ = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
151
0
def __lowerCamelCase ( snake_case__ = 10_00 ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE = 3 _SCREAMING_SNAKE_CASE = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 15 == 0: result -= a a += 1 return result if __name__ == "__main__": print(f"{solution() = }")
569
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class __UpperCAmelCase (unittest.TestCase ): @slow def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = FlaxMTaForConditionalGeneration.from_pretrained("""google/mt5-small""" ) _SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained("""google/mt5-small""" ) _SCREAMING_SNAKE_CASE = tokenizer("""Hello there""" , return_tensors="""np""" ).input_ids _SCREAMING_SNAKE_CASE = tokenizer("""Hi I am""" , return_tensors="""np""" ).input_ids _SCREAMING_SNAKE_CASE = shift_tokens_right(UpperCAmelCase_ , model.config.pad_token_id , model.config.decoder_start_token_id ) _SCREAMING_SNAKE_CASE = model(UpperCAmelCase_ , decoder_input_ids=UpperCAmelCase_ ).logits _SCREAMING_SNAKE_CASE = optax.softmax_cross_entropy(UpperCAmelCase_ , onehot(UpperCAmelCase_ , logits.shape[-1] ) ).mean() _SCREAMING_SNAKE_CASE = -(labels.shape[-1] * loss.item()) _SCREAMING_SNAKE_CASE = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
569
1
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __magic_name__ : str = abspath(join(dirname(dirname(dirname(__file__))), """src""")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="""ignore""", category=FutureWarning) def UpperCamelCase (SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(SCREAMING_SNAKE_CASE ) def UpperCamelCase (SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_terminal_summary_main UpperCamelCase : Union[str, Any] = terminalreporter.config.getoption("""--make-reports""" ) if make_reports: pytest_terminal_summary_main(SCREAMING_SNAKE_CASE , id=SCREAMING_SNAKE_CASE )
102
from argparse import ArgumentParser from .env import EnvironmentCommand def snake_case_ ( ): __lowercase = ArgumentParser("Diffusers CLI tool" , usage="diffusers-cli <command> [<args>]" ) __lowercase = parser.add_subparsers(help="diffusers-cli command helpers" ) # Register commands EnvironmentCommand.register_subcommand(_SCREAMING_SNAKE_CASE ) # Let's go __lowercase = parser.parse_args() if not hasattr(_SCREAMING_SNAKE_CASE , "func" ): parser.print_help() exit(1 ) # Run __lowercase = args.func(_SCREAMING_SNAKE_CASE ) service.run() if __name__ == "__main__": main()
402
0
def __UpperCAmelCase ( snake_case_ : int ): '''simple docstring''' return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def __UpperCAmelCase ( snake_case_ : int ): '''simple docstring''' UpperCAmelCase: str = 0 UpperCAmelCase: str = number while duplicate > 0: UpperCAmelCase: Any = divmod(snake_case_ , 1_0 ) fact_sum += factorial(snake_case_ ) return fact_sum == number if __name__ == "__main__": print('Program to check whether a number is a Krisnamurthy Number or not.') snake_case_ : List[str] = int(input('Enter number: ').strip()) print( f"""{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.""" )
702
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation snake_case_ : Tuple = logging.get_logger(__name__) snake_case_ : Tuple = {'vocab_file': 'vocab.txt', 'emoji_file': 'emoji.json'} snake_case_ : Union[str, Any] = { 'vocab_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt', }, 'emoji_file': { 'abeja/gpt-neox-japanese-2.7b': 'https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json', }, } snake_case_ : Union[str, Any] = { 'abeja/gpt-neox-japanese-2.7b': 2_0_4_8, } def __UpperCAmelCase ( snake_case_ : Any , snake_case_ : List[str] ): '''simple docstring''' with open(snake_case_ , "r" , encoding="utf-8" ) as f: UpperCAmelCase: Optional[Any] = json.loads(f.read() ) UpperCAmelCase: List[Any] = collections.OrderedDict() UpperCAmelCase: List[str] = collections.OrderedDict() UpperCAmelCase: Any = collections.OrderedDict() with open(snake_case_ , "r" , encoding="utf-8" ) as f: UpperCAmelCase: Optional[Any] = f.readlines() UpperCAmelCase: int = [[t.rstrip("\n" )] if (t == "," or "," not in t) else t.rstrip("\n" ).split("," ) for t in token] for idx, b in enumerate(snake_case_ ): UpperCAmelCase: Optional[int] = b UpperCAmelCase: int = idx for wd in b: UpperCAmelCase: str = idx return vocab, raw_vocab, ids_to_tokens, emoji class __lowerCamelCase ( lowercase ): lowerCamelCase__: Dict = VOCAB_FILES_NAMES lowerCamelCase__: List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__: Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__: Dict = ['''input_ids''', '''attention_mask'''] def __init__( self , __snake_case , __snake_case , __snake_case="<|endoftext|>" , __snake_case="<|endoftext|>" , __snake_case="<|startoftext|>" , __snake_case="<|endoftext|>" , __snake_case=False , **__snake_case , ) -> int: """simple docstring""" super().__init__( unk_token=__snake_case , pad_token=__snake_case , bos_token=__snake_case , eos_token=__snake_case , do_clean_text=__snake_case , **__snake_case , ) if not os.path.isfile(__snake_case ): raise ValueError( F'Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained' " model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`" ) if not os.path.isfile(__snake_case ): raise ValueError( F'Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google' " pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`" ) UpperCAmelCase: Optional[int] = do_clean_text UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase: str = load_vocab_and_emoji(__snake_case , __snake_case ) UpperCAmelCase: Optional[int] = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def A__ ( self ) -> Tuple: """simple docstring""" return len(self.raw_vocab ) def A__ ( self ) -> Optional[int]: """simple docstring""" return dict(self.raw_vocab , **self.added_tokens_encoder ) def A__ ( self , __snake_case ) -> int: """simple docstring""" return self.subword_tokenizer.tokenize(__snake_case , clean=self.do_clean_text ) def A__ ( self , __snake_case ) -> Any: """simple docstring""" return self.vocab.get(__snake_case , self.vocab.get(self.unk_token ) ) def A__ ( self , __snake_case ) -> Optional[int]: """simple docstring""" return self.subword_tokenizer.convert_id_to_token(__snake_case ) def A__ ( self , __snake_case ) -> Tuple: """simple docstring""" UpperCAmelCase: Any = "".join(__snake_case ).strip() return out_string def A__ ( self , __snake_case ) -> List[int]: """simple docstring""" UpperCAmelCase: Union[str, Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__snake_case , add_special_tokens=__snake_case ) + [self.eos_token_id] ) if len(__snake_case ) > self.model_max_length: UpperCAmelCase: List[Any] = input_ids[-self.model_max_length :] return input_ids def A__ ( self , __snake_case , __snake_case = None ) -> Tuple[str]: """simple docstring""" UpperCAmelCase: Tuple = 0 if os.path.isdir(__snake_case ): UpperCAmelCase: Optional[Any] = os.path.join( __snake_case , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase: Union[str, Any] = os.path.join( __snake_case , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["emoji_file"] ) else: UpperCAmelCase: Optional[int] = ( (filename_prefix + "-" if filename_prefix else "") + save_directory + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase: Any = ( (filename_prefix + "-" if filename_prefix else "") + save_directory + VOCAB_FILES_NAMES["emoji_file"] ) with open(__snake_case , "w" , encoding="utf-8" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' " Please check that the vocabulary is not corrupted!" ) UpperCAmelCase: Any = token_index writer.write(",".join(__snake_case ) + "\n" ) index += 1 with open(__snake_case , "w" , encoding="utf-8" ) as writer: json.dump(self.emoji , __snake_case ) return vocab_file, emoji_file class __lowerCamelCase ( lowercase ): def __init__( self , __snake_case , __snake_case , __snake_case ) -> str: """simple docstring""" UpperCAmelCase: List[Any] = vocab # same as swe UpperCAmelCase: Optional[Any] = ids_to_tokens # same as bpe UpperCAmelCase: Union[str, Any] = emoji UpperCAmelCase: Dict = np.max([len(__snake_case ) for w in self.vocab.keys()] ) UpperCAmelCase: Any = re.compile(R"(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)" ) UpperCAmelCase: Tuple = re.compile(R"[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*" ) UpperCAmelCase: List[Any] = re.compile(R"[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}" ) UpperCAmelCase: List[Any] = re.compile( R"([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*" ) UpperCAmelCase: Any = re.compile( R"(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*" ) UpperCAmelCase: List[str] = re.compile( R"((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*" ) UpperCAmelCase: str = "─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿" UpperCAmelCase: Any = "▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟" UpperCAmelCase: str = str.maketrans({k: "<BLOCK>" for k in keisen + blocks} ) def __len__( self ) -> List[Any]: """simple docstring""" return len(self.ids_to_tokens ) def A__ ( self , __snake_case ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase: Optional[Any] = self.content_repattera.sub("<URL>" , __snake_case ) UpperCAmelCase: str = self.content_repattera.sub("<EMAIL>" , __snake_case ) UpperCAmelCase: List[Any] = self.content_repattera.sub("<TEL>" , __snake_case ) UpperCAmelCase: Optional[int] = self.content_repattera.sub("<DATE>" , __snake_case ) UpperCAmelCase: str = self.content_repattera.sub("<DATE>" , __snake_case ) UpperCAmelCase: List[str] = self.content_repattera.sub("<PRICE>" , __snake_case ) UpperCAmelCase: Tuple = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: UpperCAmelCase: Union[str, Any] = content.replace("<BLOCK><BLOCK>" , "<BLOCK>" ) return content def A__ ( self , __snake_case , __snake_case=False ) -> List[str]: """simple docstring""" UpperCAmelCase: List[Any] = text.replace(" " , "<SP>" ) UpperCAmelCase: Tuple = text.replace(" " , "<SP>" ) UpperCAmelCase: Any = text.replace("\r\n" , "<BR>" ) UpperCAmelCase: Union[str, Any] = text.replace("\n" , "<BR>" ) UpperCAmelCase: Optional[int] = text.replace("\r" , "<BR>" ) UpperCAmelCase: int = text.replace("\t" , "<TAB>" ) UpperCAmelCase: Optional[Any] = text.replace("—" , "ー" ) UpperCAmelCase: Any = text.replace("−" , "ー" ) for k, v in self.emoji["emoji"].items(): if k in text: UpperCAmelCase: Dict = text.replace(__snake_case , __snake_case ) if clean: UpperCAmelCase: Optional[Any] = self.clean_text(__snake_case ) def check_simbol(__snake_case ): UpperCAmelCase: str = x.encode() if len(__snake_case ) == 1 and len(__snake_case ) == 2: UpperCAmelCase: str = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xC2A1 and c <= 0xC2BF) or (c >= 0xC780 and c <= 0xC783) or (c >= 0xCAB9 and c <= 0xCBBF) or (c >= 0xCC80 and c <= 0xCDA2) ): return True return False def checkuae(__snake_case ): UpperCAmelCase: int = x.encode() if len(__snake_case ) == 1 and len(__snake_case ) == 3: UpperCAmelCase: Dict = (int(e[0] ) << 1_6) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xE2_8080 and c <= 0xE2_B07F: return True return False UpperCAmelCase: int = 0 UpperCAmelCase: List[Any] = [] while pos < len(__snake_case ): UpperCAmelCase: List[str] = min(len(__snake_case ) , pos + self.maxlen + 1 ) if text[pos] == "<" else pos + 3 UpperCAmelCase: Tuple = [] # (token_id, token, pos) for e in range(__snake_case , __snake_case , -1 ): UpperCAmelCase: Any = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(__snake_case ) > 2: UpperCAmelCase: Optional[Any] = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(__snake_case ) > 0: # the smallest token_id is adopted UpperCAmelCase , UpperCAmelCase , UpperCAmelCase: Tuple = sorted(__snake_case , key=lambda __snake_case : x[0] )[0] result.append(__snake_case ) UpperCAmelCase: Dict = e else: UpperCAmelCase: Union[str, Any] = pos + 1 UpperCAmelCase: int = text[pos:end] if check_simbol(__snake_case ): result.append("<KIGOU>" ) elif checkuae(__snake_case ): result.append("<U2000U2BFF>" ) else: for i in wd.encode("utf-8" ): result.append("<|byte%d|>" % i ) UpperCAmelCase: List[Any] = end return result def A__ ( self , __snake_case , __snake_case="\n" ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase: int = [] UpperCAmelCase: int = [] UpperCAmelCase: str = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(__snake_case ) > 0: words.append(bytearray(__snake_case ).decode("utf-8" , errors="replace" ) ) UpperCAmelCase: Optional[Any] = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["emoji_inv"][word] ) elif word == "<SP>": words.append(" " ) elif word == "<BR>": words.append(__snake_case ) elif word == "<TAB>": words.append("\t" ) elif word == "<BLOCK>": words.append("▀" ) elif word == "<KIGOU>": words.append("ǀ" ) elif word == "<U2000U2BFF>": words.append("‖" ) else: words.append(__snake_case ) if len(__snake_case ) > 0: words.append(bytearray(__snake_case ).decode("utf-8" , errors="replace" ) ) UpperCAmelCase: Tuple = "".join(__snake_case ) return text
166
0
from __future__ import annotations def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : Optional[int] = sorted(numsa + numsa ) A_ , A_ : Any = divmod(len(SCREAMING_SNAKE_CASE ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = [float(x) for x in input("""Enter the elements of first array: """).split()] UpperCamelCase = [float(x) for x in input("""Enter the elements of second array: """).split()] print(F'''The median of two arrays is: {median_of_two_arrays(array_a, array_a)}''')
590
UpperCamelCase = 256 # Modulus to hash a string UpperCamelCase = 100_0003 def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : Any = len(SCREAMING_SNAKE_CASE ) A_ : int = len(SCREAMING_SNAKE_CASE ) if p_len > t_len: return False A_ : int = 0 A_ : Dict = 0 A_ : Optional[int] = 1 # Calculating the hash of pattern and substring of text for i in range(SCREAMING_SNAKE_CASE ): A_ : List[Any] = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus A_ : Dict = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue A_ : Tuple = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash A_ : List[str] = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def _SCREAMING_SNAKE_CASE ( ): A_ : List[Any] = '''abc1abc12''' A_ : str = '''alskfjaldsabc1abc1abc12k23adsfabcabc''' A_ : Optional[Any] = '''alskfjaldsk23adsfabcabc''' assert rabin_karp(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) and not rabin_karp(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Test 2) A_ : List[str] = '''ABABX''' A_ : Tuple = '''ABABZABABYABABX''' assert rabin_karp(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Test 3) A_ : Optional[int] = '''AAAB''' A_ : Optional[Any] = '''ABAAAAAB''' assert rabin_karp(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Test 4) A_ : Optional[Any] = '''abcdabcy''' A_ : str = '''abcxabcdabxabcdabcdabcy''' assert rabin_karp(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Test 5) A_ : Tuple = '''Lü''' A_ : Dict = '''Lüsai''' assert rabin_karp(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) A_ : Dict = '''Lue''' assert not rabin_karp(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print('''Success.''' ) if __name__ == "__main__": test_rabin_karp()
590
1
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) __UpperCAmelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class SCREAMING_SNAKE_CASE : '''simple docstring''' __UpperCamelCase = field( default=snake_case , metadata={"help": "Model type selected in the list: " + ", ".join(snake_case )} ) __UpperCamelCase = field( default=snake_case , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) __UpperCamelCase = field( default=128 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) __UpperCamelCase = field( default=128 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) __UpperCamelCase = field( default=64 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) __UpperCamelCase = field( default=30 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) __UpperCamelCase = field( default=snake_case , metadata={"help": "Overwrite the cached training and evaluation sets"} ) __UpperCamelCase = field( default=snake_case , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) __UpperCamelCase = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) __UpperCamelCase = field( default=20 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) __UpperCamelCase = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) __UpperCamelCase = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class SCREAMING_SNAKE_CASE ( snake_case ): '''simple docstring''' __UpperCamelCase = "train" __UpperCamelCase = "dev" class SCREAMING_SNAKE_CASE ( snake_case ): '''simple docstring''' __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 def __init__( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = Split.train , SCREAMING_SNAKE_CASE__ = False , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = "pt" , ): '''simple docstring''' snake_case: List[str] = args snake_case: int = is_language_sensitive snake_case: List[str] = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): try: snake_case: Union[str, Any] = Split[mode] except KeyError: raise KeyError('mode is not a valid split name' ) snake_case: Optional[Any] = mode # Load data features from cache or dataset file snake_case: Union[str, Any] = 'v2' if args.version_2_with_negative else 'v1' snake_case: Union[str, Any] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}""" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. snake_case: Optional[int] = cached_features_file + '.lock' with FileLock(SCREAMING_SNAKE_CASE__ ): if os.path.exists(SCREAMING_SNAKE_CASE__ ) and not args.overwrite_cache: snake_case: Any = time.time() snake_case: Dict = torch.load(SCREAMING_SNAKE_CASE__ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. snake_case: int = self.old_features['features'] snake_case: Any = self.old_features.get('dataset' , SCREAMING_SNAKE_CASE__ ) snake_case: Optional[Any] = self.old_features.get('examples' , SCREAMING_SNAKE_CASE__ ) logger.info( F"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( F"""Deleting cached file {cached_features_file} will allow dataset and examples to be cached in""" ' future run' ) else: if mode == Split.dev: snake_case: Any = self.processor.get_dev_examples(args.data_dir ) else: snake_case: Optional[int] = self.processor.get_train_examples(args.data_dir ) snake_case , snake_case: Optional[Any] = squad_convert_examples_to_features( examples=self.examples , tokenizer=SCREAMING_SNAKE_CASE__ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=SCREAMING_SNAKE_CASE__ , ) snake_case: List[Any] = time.time() torch.save( {'features': self.features, 'dataset': self.dataset, 'examples': self.examples} , SCREAMING_SNAKE_CASE__ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self ): '''simple docstring''' return len(self.features ) def __getitem__( self , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' snake_case: Dict = self.features[i] snake_case: Optional[Any] = torch.tensor(feature.input_ids , dtype=torch.long ) snake_case: int = torch.tensor(feature.attention_mask , dtype=torch.long ) snake_case: Any = torch.tensor(feature.token_type_ids , dtype=torch.long ) snake_case: Union[str, Any] = torch.tensor(feature.cls_index , dtype=torch.long ) snake_case: Union[str, Any] = torch.tensor(feature.p_mask , dtype=torch.float ) snake_case: List[Any] = torch.tensor(feature.is_impossible , dtype=torch.float ) snake_case: int = { 'input_ids': input_ids, 'attention_mask': attention_mask, 'token_type_ids': token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({'cls_index': cls_index, 'p_mask': p_mask} ) if self.args.version_2_with_negative: inputs.update({'is_impossible': is_impossible} ) if self.is_language_sensitive: inputs.update({'langs': (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: snake_case: str = torch.tensor(feature.start_position , dtype=torch.long ) snake_case: str = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({'start_positions': start_positions, 'end_positions': end_positions} ) return inputs
692
'''simple docstring''' import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } __UpperCAmelCase = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", ] def lowerCAmelCase_ ( __A : Any , __A : Optional[Any] , __A : Union[str, Any] , __A : int , __A : Optional[int] ): '''simple docstring''' for attribute in key.split('.' ): snake_case: List[str] = getattr(__A , __A ) if weight_type is not None: snake_case: Optional[int] = getattr(__A , __A ).shape else: snake_case: Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": snake_case: Optional[int] = value elif weight_type == "weight_g": snake_case: List[str] = value elif weight_type == "weight_v": snake_case: Dict = value elif weight_type == "bias": snake_case: Optional[Any] = value else: snake_case: int = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCAmelCase_ ( __A : List[Any] , __A : List[str] ): '''simple docstring''' snake_case: List[Any] = [] snake_case: List[Any] = fairseq_model.state_dict() snake_case: Union[str, Any] = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight snake_case: Dict = None for name, value in fairseq_dict.items(): snake_case: Tuple = False if "conv_layers" in name: load_conv_layer( __A , __A , __A , __A , hf_model.config.feat_extract_norm == 'group' , ) snake_case: List[Any] = True elif name.split('.' )[0] == "proj": snake_case: List[Any] = fairseq_model.proj snake_case: int = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: snake_case: int = True if "*" in mapped_key: snake_case: List[str] = name.split(__A )[0].split('.' )[-2] snake_case: Dict = mapped_key.replace('*' , __A ) if "weight_g" in name: snake_case: Tuple = 'weight_g' elif "weight_v" in name: snake_case: int = 'weight_v' elif "bias" in name: snake_case: Tuple = 'bias' elif "weight" in name: snake_case: List[Any] = 'weight' else: snake_case: Any = None set_recursively(__A , __A , __A , __A , __A ) continue if not is_used: unused_weights.append(__A ) logger.warning(f"""Unused weights: {unused_weights}""" ) return proj_weight def lowerCAmelCase_ ( __A : List[str] , __A : List[Any] , __A : int , __A : Optional[Any] , __A : Optional[int] ): '''simple docstring''' snake_case: int = full_name.split('conv_layers.' )[-1] snake_case: Tuple = name.split('.' ) snake_case: Any = int(items[0] ) snake_case: Optional[int] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) snake_case: Tuple = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) snake_case: int = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) snake_case: Any = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) snake_case: str = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(__A ) def lowerCAmelCase_ ( __A : Dict ): '''simple docstring''' snake_case , snake_case: List[Any] = emb.weight.shape snake_case: Optional[int] = nn.Linear(__A , __A , bias=__A ) snake_case: Any = emb.weight.data return lin_layer def lowerCAmelCase_ ( __A : Optional[int] ): '''simple docstring''' with open(__A , 'r' , encoding='utf-8' ) as f: snake_case: List[Any] = f.readlines() snake_case: Any = [line.split(' ' )[0] for line in lines] snake_case: int = len(__A ) snake_case: Dict = { '<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3, } vocab_dict.update(dict(zip(__A , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def lowerCAmelCase_ ( __A : Optional[Any] , __A : Union[str, Any] , __A : Dict , __A : Any , __A : List[Any] , __A : int , __A : str , ): '''simple docstring''' snake_case: Union[str, Any] = WavaVecaConfig.from_pretrained(__A ) snake_case: str = SpeechaTextaConfig.from_pretrained( __A , vocab_size=__A , decoder_layers=__A , do_stable_layer_norm=__A ) snake_case: List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=__A , return_attention_mask=__A , ) snake_case , snake_case , snake_case: List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) snake_case: List[Any] = model[0].eval() # set weights for wav2vec2 encoder snake_case: Optional[Any] = WavaVecaModel(__A ) snake_case: Any = recursively_load_weights_wavaveca(model.encoder , __A ) snake_case: Union[str, Any] = SpeechaTextaForCausalLM(__A ) snake_case , snake_case: Optional[Any] = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=__A ) # set output linear layer unexpected_keys.remove('embed_out' ) snake_case: str = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(f"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(f"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) snake_case: int = SpeechEncoderDecoderModel(encoder=__A , decoder=__A ) snake_case: List[Any] = False # add projection layer snake_case: Union[str, Any] = nn.Parameter(projection_layer.weight ) snake_case: Union[str, Any] = nn.Parameter(projection_layer.bias ) snake_case: List[Any] = create_vocab_dict(__A ) with open(os.path.join(__A , 'vocab.json' ) , 'w' ) as fp: json.dump(__A , __A ) snake_case: Union[str, Any] = SpeechaTextaTokenizer(os.path.join(__A , 'vocab.json' ) ) tokenizer.save_pretrained(__A ) snake_case: Tuple = hf_wavavec.config.to_dict() snake_case: int = tokenizer.pad_token_id snake_case: Dict = tokenizer.bos_token_id snake_case: Optional[int] = tokenizer.eos_token_id snake_case: Dict = 'speech_to_text_2' snake_case: Optional[Any] = 'wav2vec2' snake_case: Tuple = SpeechEncoderDecoderConfig.from_dict(__A ) hf_wavavec.save_pretrained(__A ) feature_extractor.save_pretrained(__A ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument( "--encoder_config_path", default="facebook/wav2vec2-large-lv60", type=str, help="Path to hf encoder wav2vec2 checkpoint config", ) parser.add_argument( "--decoder_config_path", default="facebook/s2t-small-mustc-en-fr-st", type=str, help="Path to hf decoder s2t checkpoint config", ) parser.add_argument("--vocab_size", default=10_224, type=int, help="Vocab size of decoder") parser.add_argument("--num_decoder_layers", default=7, type=int, help="Number of decoder layers") __UpperCAmelCase = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
692
1
import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class lowercase ( UpperCamelCase__ ): def __snake_case( self : List[Any] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def __snake_case( self : int ) -> List[str]: '''simple docstring''' with self.assertRaises(__snake_case ): SCREAMING_SNAKE_CASE = pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def __snake_case( self : Tuple ) -> Tuple: '''simple docstring''' with self.assertRaises(__snake_case ): SCREAMING_SNAKE_CASE = pa.array(TypedSequence([1, 2, 3] , try_type=Value("bool" ) , type=Value("int64" ) ) ) def __snake_case( self : Union[str, Any] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = pa.array(TypedSequence([1, 2, 3] , type=Value("int32" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def __snake_case( self : str ) -> Optional[int]: '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): SCREAMING_SNAKE_CASE = pa.array(TypedSequence(["foo", "bar"] , type=Value("int64" ) ) ) def __snake_case( self : List[str] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = pa.array(TypedSequence([1, 2, 3] , try_type=Value("int32" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def __snake_case( self : List[str] ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = pa.array(TypedSequence(["foo", "bar"] , try_type=Value("int64" ) ) ) self.assertEqual(arr.type , pa.string() ) def __snake_case( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , "int64" ) ) def __snake_case( self : int ) -> Dict: '''simple docstring''' with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): SCREAMING_SNAKE_CASE = pa.array(TypedSequence(["foo", "bar"] , type=ArrayaD((1, 3) , "int64" ) ) ) def __snake_case( self : int ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , "int64" ) ) def __snake_case( self : Union[str, Any] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = pa.array(TypedSequence(["foo", "bar"] , try_type=ArrayaD((1, 3) , "int64" ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def __snake_case( self : Optional[Any] ) -> Dict: '''simple docstring''' import PIL.Image SCREAMING_SNAKE_CASE = PIL.Image.fromarray(np.arange(10 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( "datasets.arrow_writer.cast_to_python_objects" , side_effect=__snake_case ) as mock_cast_to_python_objects: SCREAMING_SNAKE_CASE = pa.array(TypedSequence([{"path": None, "bytes": b"image_bytes"}, pil_image] , type=Image() ) ) SCREAMING_SNAKE_CASE = mock_cast_to_python_objects.call_args_list[-1] self.assertIn("optimize_list_casting" , __snake_case ) self.assertFalse(kwargs["optimize_list_casting"] ) def __lowerCamelCase (UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[str, Any] ): SCREAMING_SNAKE_CASE = pa.BufferReader(UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , pa.Buffer ) else pa.memory_map(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = pa.ipc.open_stream(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize("writer_batch_size" , [None, 1, 1_0] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase (UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] ): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() SCREAMING_SNAKE_CASE = pa.schema(UpperCAmelCase__ ) if fields else None with ArrowWriter(stream=UpperCAmelCase__ , schema=UpperCAmelCase__ , writer_batch_size=UpperCAmelCase__ ) as writer: writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: SCREAMING_SNAKE_CASE = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(UpperCAmelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def __lowerCamelCase (): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() SCREAMING_SNAKE_CASE = Features({"labels": ClassLabel(names=["neg", "pos"] )} ) with ArrowWriter(stream=UpperCAmelCase__ , features=UpperCAmelCase__ ) as writer: writer.write({"labels": 0} ) writer.write({"labels": 1} ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata SCREAMING_SNAKE_CASE = pa.BufferReader(output.getvalue() ) SCREAMING_SNAKE_CASE = pa.ipc.open_stream(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = f.read_all() SCREAMING_SNAKE_CASE = pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(UpperCAmelCase__ ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 1_0] ) def __lowerCamelCase (UpperCAmelCase__ : Optional[int] ): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() with ArrowWriter( stream=UpperCAmelCase__ , writer_batch_size=UpperCAmelCase__ , hash_salt="split_name" , check_duplicates=UpperCAmelCase__ , ) as writer: with pytest.raises(UpperCAmelCase__ ): writer.write({"col_1": "foo", "col_2": 1} , key=[1, 2] ) SCREAMING_SNAKE_CASE = writer.finalize() @pytest.mark.parametrize("writer_batch_size" , [None, 2, 1_0] ) def __lowerCamelCase (UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() with ArrowWriter( stream=UpperCAmelCase__ , writer_batch_size=UpperCAmelCase__ , hash_salt="split_name" , check_duplicates=UpperCAmelCase__ , ) as writer: with pytest.raises(UpperCAmelCase__ ): writer.write({"col_1": "foo", "col_2": 1} , key=1_0 ) writer.write({"col_1": "bar", "col_2": 2} , key=1_0 ) SCREAMING_SNAKE_CASE = writer.finalize() @pytest.mark.parametrize("writer_batch_size" , [None, 2, 1_0] ) def __lowerCamelCase (UpperCAmelCase__ : List[Any] ): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() with ArrowWriter( stream=UpperCAmelCase__ , writer_batch_size=UpperCAmelCase__ , hash_salt="split_name" , check_duplicates=UpperCAmelCase__ , ) as writer: writer.write({"col_1": "foo", "col_2": 1} , key=1 ) writer.write({"col_1": "bar", "col_2": 2} , key=2 ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 1_0] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase (UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() SCREAMING_SNAKE_CASE = pa.schema(UpperCAmelCase__ ) if fields else None with ArrowWriter(stream=UpperCAmelCase__ , schema=UpperCAmelCase__ , writer_batch_size=UpperCAmelCase__ ) as writer: writer.write_batch({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) writer.write_batch({"col_1": [], "col_2": []} ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: SCREAMING_SNAKE_CASE = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(UpperCAmelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 1_0] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase (UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict ): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() SCREAMING_SNAKE_CASE = pa.schema(UpperCAmelCase__ ) if fields else None with ArrowWriter(stream=UpperCAmelCase__ , schema=UpperCAmelCase__ , writer_batch_size=UpperCAmelCase__ ) as writer: writer.write_table(pa.Table.from_pydict({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: SCREAMING_SNAKE_CASE = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(UpperCAmelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("writer_batch_size" , [None, 1, 1_0] ) @pytest.mark.parametrize( "fields" , [None, {"col_1": pa.string(), "col_2": pa.intaa()}, {"col_1": pa.string(), "col_2": pa.intaa()}] ) def __lowerCamelCase (UpperCAmelCase__ : Dict , UpperCAmelCase__ : Any ): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() SCREAMING_SNAKE_CASE = pa.schema(UpperCAmelCase__ ) if fields else None with ArrowWriter(stream=UpperCAmelCase__ , schema=UpperCAmelCase__ , writer_batch_size=UpperCAmelCase__ ) as writer: writer.write_row(pa.Table.from_pydict({"col_1": ["foo"], "col_2": [1]} ) ) writer.write_row(pa.Table.from_pydict({"col_1": ["bar"], "col_2": [2]} ) ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: SCREAMING_SNAKE_CASE = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(UpperCAmelCase__ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def __lowerCamelCase (): with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE = {"""col_1""": pa.string(), """col_2""": pa.intaa()} SCREAMING_SNAKE_CASE = os.path.join(UpperCAmelCase__ , "test.arrow" ) with ArrowWriter(path=UpperCAmelCase__ , schema=pa.schema(UpperCAmelCase__ ) ) as writer: writer.write_batch({"col_1": ["foo", "bar"], "col_2": [1, 2]} ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(UpperCAmelCase__ , metadata=writer._schema.metadata ) _check_output(UpperCAmelCase__ , 1 ) def __lowerCamelCase (UpperCAmelCase__ : List[Any] ): if pa.types.is_list(UpperCAmelCase__ ): return get_base_dtype(arr_type.value_type ) else: return arr_type def __lowerCamelCase (UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Dict ): if isinstance(lst[0] , UpperCAmelCase__ ): change_first_primitive_element_in_list(lst[0] , UpperCAmelCase__ ) else: SCREAMING_SNAKE_CASE = value @pytest.mark.parametrize("optimized_int_type, expected_dtype" , [(None, pa.intaa()), (Value("int32" ), pa.intaa())] ) @pytest.mark.parametrize("sequence" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def __lowerCamelCase (UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any ): SCREAMING_SNAKE_CASE = pa.array(TypedSequence(UpperCAmelCase__ , optimized_int_type=UpperCAmelCase__ ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( "col, expected_dtype" , [ ("attention_mask", pa.inta()), ("special_tokens_mask", pa.inta()), ("token_type_ids", pa.inta()), ("input_ids", pa.intaa()), ("other", pa.intaa()), ] , ) @pytest.mark.parametrize("sequence" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def __lowerCamelCase (UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : str ): # in range SCREAMING_SNAKE_CASE = pa.array(OptimizedTypedSequence(UpperCAmelCase__ , col=UpperCAmelCase__ ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications SCREAMING_SNAKE_CASE = copy.deepcopy(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(UpperCAmelCase__ , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = pa.array(OptimizedTypedSequence(UpperCAmelCase__ , col=UpperCAmelCase__ ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize("raise_exception" , [False, True] ) def __lowerCamelCase (UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any ): SCREAMING_SNAKE_CASE = str(tmp_path / "dataset-train.arrow" ) try: with ArrowWriter(path=UpperCAmelCase__ ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def __lowerCamelCase (UpperCAmelCase__ : Tuple ): SCREAMING_SNAKE_CASE = """mock://dataset-train.arrow""" with ArrowWriter(path=UpperCAmelCase__ , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(UpperCAmelCase__ ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(UpperCAmelCase__ ) def __lowerCamelCase (): SCREAMING_SNAKE_CASE = pa.BufferOutputStream() with ParquetWriter(stream=UpperCAmelCase__ ) as writer: writer.write({"col_1": "foo", "col_2": 1} ) writer.write({"col_1": "bar", "col_2": 2} ) SCREAMING_SNAKE_CASE = writer.finalize() assert num_examples == 2 assert num_bytes > 0 SCREAMING_SNAKE_CASE = pa.BufferReader(output.getvalue() ) SCREAMING_SNAKE_CASE = pq.read_table(UpperCAmelCase__ ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize("embed_local_files" , [False, True] ) def __lowerCamelCase (UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict ): import PIL.Image SCREAMING_SNAKE_CASE = str(tmp_path / "test_image_rgb.jpg" ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(UpperCAmelCase__ , format="png" ) SCREAMING_SNAKE_CASE = pa.BufferOutputStream() with ParquetWriter( stream=UpperCAmelCase__ , features=Features({"image": Image()} ) , embed_local_files=UpperCAmelCase__ ) as writer: writer.write({"image": image_path} ) writer.finalize() SCREAMING_SNAKE_CASE = pa.BufferReader(output.getvalue() ) SCREAMING_SNAKE_CASE = pq.read_table(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = pa_table.to_pydict() if embed_local_files: assert isinstance(out["image"][0]["path"] , UpperCAmelCase__ ) with open(UpperCAmelCase__ , "rb" ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def __lowerCamelCase (): SCREAMING_SNAKE_CASE = pa.schema([pa.field("col_1" , pa.string() , nullable=UpperCAmelCase__ )] ) SCREAMING_SNAKE_CASE = pa.BufferOutputStream() with ArrowWriter(stream=UpperCAmelCase__ ) as writer: writer._build_writer(inferred_schema=UpperCAmelCase__ ) assert writer._schema == pa.schema([pa.field("col_1" , pa.string() )] )
403
import heapq def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : list[list] =[] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(lowerCamelCase , [-1 * len(lowerCamelCase ), (key, value)] ) # chosen_vertices = set of chosen vertices __magic_name__ : Tuple =set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices __magic_name__ : Tuple =heapq.heappop(lowerCamelCase )[1][0] chosen_vertices.add(lowerCamelCase ) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: __magic_name__ : Tuple =elem[1][1].index(lowerCamelCase ) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(lowerCamelCase ) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ : Optional[int] = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(F"""Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}""")
21
0
import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class A (ctypes.Structure ): '''simple docstring''' __lowerCamelCase : Dict = [('''size''', ctypes.c_int), ('''visible''', ctypes.c_byte)] def __lowerCamelCase ( ) -> List[Any]: """simple docstring""" if os.name == "nt": A__ = CursorInfo() A__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__a , ctypes.byref(__a ) ) A__ = False ctypes.windll.kernelaa.SetConsoleCursorInfo(__a , ctypes.byref(__a ) ) elif os.name == "posix": sys.stdout.write("""\033[?25l""" ) sys.stdout.flush() def __lowerCamelCase ( ) -> Dict: """simple docstring""" if os.name == "nt": A__ = CursorInfo() A__ = ctypes.windll.kernelaa.GetStdHandle(-1_1 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__a , ctypes.byref(__a ) ) A__ = True ctypes.windll.kernelaa.SetConsoleCursorInfo(__a , ctypes.byref(__a ) ) elif os.name == "posix": sys.stdout.write("""\033[?25h""" ) sys.stdout.flush() @contextmanager def __lowerCamelCase ( ) -> List[str]: """simple docstring""" try: hide_cursor() yield finally: show_cursor()
247
from __future__ import annotations A : List[Any] = 1_0 def __lowerCamelCase ( __a :list[int] ) -> list[int]: """simple docstring""" A__ = 1 A__ = max(__a ) while placement <= max_digit: # declare and initialize empty buckets A__ = [[] for _ in range(__a )] # split list_of_ints between the buckets for i in list_of_ints: A__ = int((i / placement) % RADIX ) buckets[tmp].append(__a ) # put each buckets' contents into list_of_ints A__ = 0 for b in range(__a ): for i in buckets[b]: A__ = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
247
1
"""simple docstring""" from __future__ import annotations def lowercase_ ( __UpperCAmelCase ) -> None: create_state_space_tree(__UpperCAmelCase , [] , 0 , [0 for i in range(len(__UpperCAmelCase ) )] ) def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) -> None: if index == len(__UpperCAmelCase ): print(__UpperCAmelCase ) return for i in range(len(__UpperCAmelCase ) ): if not index_used[i]: current_sequence.append(sequence[i] ) lowerCAmelCase__ : str = True create_state_space_tree(__UpperCAmelCase , __UpperCAmelCase , index + 1 , __UpperCAmelCase ) current_sequence.pop() lowerCAmelCase__ : Dict = False _A = [3, 1, 2, 4] generate_all_permutations(sequence) _A = ["A", "B", "C"] generate_all_permutations(sequence_a)
299
"""simple docstring""" from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar _A = TypeVar("""T""") _A = TypeVar("""U""") class _lowerCamelCase ( Generic[T, U] ): def __init__( self : List[Any] , UpperCamelCase : T | None , UpperCamelCase : U | None ) -> str: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = key lowerCAmelCase__ : Union[str, Any] = val lowerCAmelCase__ : DoubleLinkedListNode[T, U] | None = None lowerCAmelCase__ : DoubleLinkedListNode[T, U] | None = None def __repr__( self : List[str] ) -> str: """simple docstring""" return ( f"""Node: key: {self.key}, val: {self.val}, """ f"""has next: {bool(self.next )}, has prev: {bool(self.prev )}""" ) class _lowerCamelCase ( Generic[T, U] ): def __init__( self : int ) -> None: """simple docstring""" lowerCAmelCase__ : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.rear, self.head def __repr__( self : str ) -> str: """simple docstring""" lowerCAmelCase__ : str = ["""DoubleLinkedList"""] lowerCAmelCase__ : int = self.head while node.next is not None: rep.append(str(UpperCamelCase ) ) lowerCAmelCase__ : Optional[Any] = node.next rep.append(str(self.rear ) ) return ",\n ".join(UpperCamelCase ) def _lowerCAmelCase ( self : List[Any] , UpperCamelCase : DoubleLinkedListNode[T, U] ) -> None: """simple docstring""" lowerCAmelCase__ : List[str] = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None lowerCAmelCase__ : List[str] = node lowerCAmelCase__ : int = previous lowerCAmelCase__ : Optional[int] = node lowerCAmelCase__ : Dict = self.rear def _lowerCAmelCase ( self : List[Any] , UpperCamelCase : DoubleLinkedListNode[T, U] ) -> DoubleLinkedListNode[T, U] | None: """simple docstring""" if node.prev is None or node.next is None: return None lowerCAmelCase__ : int = node.next lowerCAmelCase__ : str = node.prev lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : List[Any] = None return node class _lowerCamelCase ( Generic[T, U] ): _lowerCamelCase :dict[Callable[[T], U], LRUCache[T, U]] = {} def __init__( self : Dict , UpperCamelCase : int ) -> str: """simple docstring""" lowerCAmelCase__ : DoubleLinkedList[T, U] = DoubleLinkedList() lowerCAmelCase__ : Tuple = capacity lowerCAmelCase__ : str = 0 lowerCAmelCase__ : Optional[int] = 0 lowerCAmelCase__ : List[Any] = 0 lowerCAmelCase__ : dict[T, DoubleLinkedListNode[T, U]] = {} def __repr__( self : Optional[Any] ) -> str: """simple docstring""" return ( f"""CacheInfo(hits={self.hits}, misses={self.miss}, """ f"""capacity={self.capacity}, current size={self.num_keys})""" ) def __contains__( self : Optional[Any] , UpperCamelCase : T ) -> bool: """simple docstring""" return key in self.cache def _lowerCAmelCase ( self : Any , UpperCamelCase : T ) -> U | None: """simple docstring""" # Note: pythonic interface would throw KeyError rather than return None if key in self.cache: self.hits += 1 lowerCAmelCase__ : DoubleLinkedListNode[T, U] = self.cache[key] lowerCAmelCase__ : Any = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(UpperCamelCase ) return node.val self.miss += 1 return None def _lowerCAmelCase ( self : Optional[Any] , UpperCamelCase : T , UpperCamelCase : U ) -> None: """simple docstring""" if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity lowerCAmelCase__ : List[Any] = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(UpperCamelCase ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 lowerCAmelCase__ : Union[str, Any] = DoubleLinkedListNode(UpperCamelCase , UpperCamelCase ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value lowerCAmelCase__ : Dict = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list lowerCAmelCase__ : Any = value self.list.add(UpperCamelCase ) @classmethod def _lowerCAmelCase ( cls : Any , UpperCamelCase : int = 1_28 ) -> Callable[[Callable[[T], U]], Callable[..., U]]: """simple docstring""" def cache_decorator_inner(UpperCamelCase : Callable[[T], U] ) -> Callable[..., U]: def cache_decorator_wrapper(*UpperCamelCase : T ) -> U: if func not in cls.decorator_function_to_instance_map: lowerCAmelCase__ : List[Any] = LRUCache(UpperCamelCase ) lowerCAmelCase__ : Dict = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: lowerCAmelCase__ : str = func(*UpperCamelCase ) cls.decorator_function_to_instance_map[func].put(args[0] , UpperCamelCase ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(UpperCamelCase , """cache_info""" , UpperCamelCase ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
299
1
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig UpperCamelCase__ = [ 'openmmlab/upernet-convnext-tiny', # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring UpperCamelCase__ = 'UperNetConfig' class a ( nn.Module ): def __init__( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = 0 , UpperCamelCase_ = False , UpperCamelCase_ = 1 , ): super().__init__() UpperCAmelCase__ : Union[str, Any] = nn.Convad( in_channels=UpperCamelCase_ , out_channels=UpperCamelCase_ , kernel_size=UpperCamelCase_ , padding=UpperCamelCase_ , bias=UpperCamelCase_ , dilation=UpperCamelCase_ , ) UpperCAmelCase__ : int = nn.BatchNormad(UpperCamelCase_ ) UpperCAmelCase__ : str = nn.ReLU() def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : List[str] = self.conv(UpperCamelCase_ ) UpperCAmelCase__ : Any = self.batch_norm(UpperCamelCase_ ) UpperCAmelCase__ : List[Any] = self.activation(UpperCamelCase_ ) return output class a ( nn.Module ): def __init__( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): super().__init__() UpperCAmelCase__ : int = [ nn.AdaptiveAvgPoolad(UpperCamelCase_ ), UperNetConvModule(UpperCamelCase_ , UpperCamelCase_ , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(UpperCamelCase_ ) , UpperCamelCase_ ) def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : str = input for layer in self.layers: UpperCAmelCase__ : Any = layer(UpperCamelCase_ ) return hidden_state class a ( nn.Module ): def __init__( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): super().__init__() UpperCAmelCase__ : Tuple = pool_scales UpperCAmelCase__ : str = align_corners UpperCAmelCase__ : Optional[Any] = in_channels UpperCAmelCase__ : str = channels UpperCAmelCase__ : List[str] = [] for i, pool_scale in enumerate(UpperCamelCase_ ): UpperCAmelCase__ : int = UperNetPyramidPoolingBlock(pool_scale=UpperCamelCase_ , in_channels=UpperCamelCase_ , channels=UpperCamelCase_ ) self.blocks.append(UpperCamelCase_ ) self.add_module(str(UpperCamelCase_ ) , UpperCamelCase_ ) def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : Tuple = [] for ppm in self.blocks: UpperCAmelCase__ : int = ppm(UpperCamelCase_ ) UpperCAmelCase__ : Any = nn.functional.interpolate( UpperCamelCase_ , size=x.size()[2:] , mode='bilinear' , align_corners=self.align_corners ) ppm_outs.append(UpperCamelCase_ ) return ppm_outs class a ( nn.Module ): def __init__( self , UpperCamelCase_ , UpperCamelCase_ ): super().__init__() UpperCAmelCase__ : Dict = config UpperCAmelCase__ : int = config.pool_scales # e.g. (1, 2, 3, 6) UpperCAmelCase__ : List[Any] = in_channels UpperCAmelCase__ : int = config.hidden_size UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : List[str] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module UpperCAmelCase__ : int = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) UpperCAmelCase__ : str = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module UpperCAmelCase__ : Any = nn.ModuleList() UpperCAmelCase__ : Union[str, Any] = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer UpperCAmelCase__ : Dict = UperNetConvModule(UpperCamelCase_ , self.channels , kernel_size=1 ) UpperCAmelCase__ : List[Any] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(UpperCamelCase_ ) self.fpn_convs.append(UpperCamelCase_ ) UpperCAmelCase__ : Optional[Any] = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self ): self.apply(self._init_weights ) def __snake_case ( self , UpperCamelCase_ ): if isinstance(UpperCamelCase_ , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : Optional[Any] = inputs[-1] UpperCAmelCase__ : Optional[Any] = [x] psp_outs.extend(self.psp_modules(UpperCamelCase_ ) ) UpperCAmelCase__ : Dict = torch.cat(UpperCamelCase_ , dim=1 ) UpperCAmelCase__ : List[Any] = self.bottleneck(UpperCamelCase_ ) return output def __snake_case ( self , UpperCamelCase_ ): # build laterals UpperCAmelCase__ : str = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(UpperCamelCase_ ) ) # build top-down path UpperCAmelCase__ : Dict = len(UpperCamelCase_ ) for i in range(used_backbone_levels - 1 , 0 , -1 ): UpperCAmelCase__ : Any = laterals[i - 1].shape[2:] UpperCAmelCase__ : List[str] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=UpperCamelCase_ , mode='bilinear' , align_corners=self.align_corners ) # build outputs UpperCAmelCase__ : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): UpperCAmelCase__ : Dict = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode='bilinear' , align_corners=self.align_corners ) UpperCAmelCase__ : List[str] = torch.cat(UpperCamelCase_ , dim=1 ) UpperCAmelCase__ : Optional[int] = self.fpn_bottleneck(UpperCamelCase_ ) UpperCAmelCase__ : Union[str, Any] = self.classifier(UpperCamelCase_ ) return output class a ( nn.Module ): def __init__( self , UpperCamelCase_ , UpperCamelCase_ = 2 , UpperCamelCase_ = 3 , UpperCamelCase_ = 1 ): super().__init__() UpperCAmelCase__ : Union[str, Any] = config UpperCAmelCase__ : Dict = config.auxiliary_in_channels UpperCAmelCase__ : str = config.auxiliary_channels UpperCAmelCase__ : Dict = config.auxiliary_num_convs UpperCAmelCase__ : Optional[int] = config.auxiliary_concat_input UpperCAmelCase__ : Optional[Any] = in_index UpperCAmelCase__ : Tuple = (kernel_size // 2) * dilation UpperCAmelCase__ : Dict = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=UpperCamelCase_ , padding=UpperCamelCase_ , dilation=UpperCamelCase_ ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=UpperCamelCase_ , padding=UpperCamelCase_ , dilation=UpperCamelCase_ ) ) if self.num_convs == 0: UpperCAmelCase__ : Union[str, Any] = nn.Identity() else: UpperCAmelCase__ : List[str] = nn.Sequential(*UpperCamelCase_ ) if self.concat_input: UpperCAmelCase__ : Tuple = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=UpperCamelCase_ , padding=kernel_size // 2 ) UpperCAmelCase__ : int = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self ): self.apply(self._init_weights ) def __snake_case ( self , UpperCamelCase_ ): if isinstance(UpperCamelCase_ , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self , UpperCamelCase_ ): # just take the relevant feature maps UpperCAmelCase__ : int = encoder_hidden_states[self.in_index] UpperCAmelCase__ : List[str] = self.convs(UpperCamelCase_ ) if self.concat_input: UpperCAmelCase__ : str = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) UpperCAmelCase__ : str = self.classifier(UpperCamelCase_ ) return output class a ( lowercase ): UpperCamelCase : List[str] = UperNetConfig UpperCamelCase : Union[str, Any] = """pixel_values""" UpperCamelCase : List[str] = True def __snake_case ( self , UpperCamelCase_ ): if isinstance(UpperCamelCase_ , UpperCamelCase_ ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self ): self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self , UpperCamelCase_ , UpperCamelCase_=False ): if isinstance(UpperCamelCase_ , UpperCamelCase_ ): UpperCAmelCase__ : List[str] = value UpperCamelCase__ = r'\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' UpperCamelCase__ = r'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( """UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes.""" , lowercase , ) class a ( lowercase ): def __init__( self , UpperCamelCase_ ): super().__init__(UpperCamelCase_ ) UpperCAmelCase__ : Optional[Any] = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) UpperCAmelCase__ : int = UperNetHead(UpperCamelCase_ , in_channels=self.backbone.channels ) UpperCAmelCase__ : List[str] = UperNetFCNHead(UpperCamelCase_ ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format('batch_size, sequence_length' ) ) @replace_return_docstrings(output_type=UpperCamelCase_ , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , ): UpperCAmelCase__ : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCAmelCase__ : Optional[int] = output_attentions if output_attentions is not None else self.config.output_attentions UpperCAmelCase__ : List[str] = self.backbone.forward_with_filtered_kwargs( UpperCamelCase_ , output_hidden_states=UpperCamelCase_ , output_attentions=UpperCamelCase_ ) UpperCAmelCase__ : Optional[Any] = outputs.feature_maps UpperCAmelCase__ : str = self.decode_head(UpperCamelCase_ ) UpperCAmelCase__ : List[str] = nn.functional.interpolate(UpperCamelCase_ , size=pixel_values.shape[2:] , mode='bilinear' , align_corners=UpperCamelCase_ ) UpperCAmelCase__ : List[Any] = None if self.auxiliary_head is not None: UpperCAmelCase__ : Tuple = self.auxiliary_head(UpperCamelCase_ ) UpperCAmelCase__ : Optional[int] = nn.functional.interpolate( UpperCamelCase_ , size=pixel_values.shape[2:] , mode='bilinear' , align_corners=UpperCamelCase_ ) UpperCAmelCase__ : List[Any] = None if labels is not None: if self.config.num_labels == 1: raise ValueError('The number of labels should be greater than one' ) else: # compute weighted loss UpperCAmelCase__ : Optional[int] = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) UpperCAmelCase__ : Union[str, Any] = loss_fct(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ : Any = loss_fct(UpperCamelCase_ , UpperCamelCase_ ) UpperCAmelCase__ : Union[str, Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: UpperCAmelCase__ : Union[str, Any] = (logits,) + outputs[1:] else: UpperCAmelCase__ : Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=UpperCamelCase_ , logits=UpperCamelCase_ , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
254
"""simple docstring""" from __future__ import annotations import math import random from typing import Any class a : def __init__( self ): UpperCAmelCase__ : list[Any] = [] UpperCAmelCase__ : int = 0 UpperCAmelCase__ : int = 0 def __snake_case ( self ): return self.head == self.tail def __snake_case ( self , UpperCamelCase_ ): self.data.append(UpperCamelCase_ ) UpperCAmelCase__ : Optional[Any] = self.tail + 1 def __snake_case ( self ): UpperCAmelCase__ : Dict = self.data[self.head] UpperCAmelCase__ : Any = self.head + 1 return ret def __snake_case ( self ): return self.tail - self.head def __snake_case ( self ): print(self.data ) print('**************' ) print(self.data[self.head : self.tail] ) class a : def __init__( self , UpperCamelCase_ ): UpperCAmelCase__ : List[Any] = data UpperCAmelCase__ : MyNode | None = None UpperCAmelCase__ : MyNode | None = None UpperCAmelCase__ : int = 1 def __snake_case ( self ): return self.data def __snake_case ( self ): return self.left def __snake_case ( self ): return self.right def __snake_case ( self ): return self.height def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : Dict = data def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : Any = node def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : Optional[Any] = node def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : Any = height def lowerCamelCase ( _snake_case ): if node is None: return 0 return node.get_height() def lowerCamelCase ( _snake_case ,_snake_case ): if a > b: return a return b def lowerCamelCase ( _snake_case ): print('left rotation node:' ,node.get_data() ) UpperCAmelCase__ : Dict = node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(_snake_case ) UpperCAmelCase__ : str = my_max(get_height(node.get_right() ) ,get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) UpperCAmelCase__ : str = my_max(get_height(ret.get_right() ) ,get_height(ret.get_left() ) ) + 1 ret.set_height(_snake_case ) return ret def lowerCamelCase ( _snake_case ): print('right rotation node:' ,node.get_data() ) UpperCAmelCase__ : List[str] = node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(_snake_case ) UpperCAmelCase__ : Dict = my_max(get_height(node.get_right() ) ,get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) UpperCAmelCase__ : Tuple = my_max(get_height(ret.get_right() ) ,get_height(ret.get_left() ) ) + 1 ret.set_height(_snake_case ) return ret def lowerCamelCase ( _snake_case ): UpperCAmelCase__ : Tuple = node.get_left() assert left_child is not None node.set_left(left_rotation(_snake_case ) ) return right_rotation(_snake_case ) def lowerCamelCase ( _snake_case ): UpperCAmelCase__ : Dict = node.get_right() assert right_child is not None node.set_right(right_rotation(_snake_case ) ) return left_rotation(_snake_case ) def lowerCamelCase ( _snake_case ,_snake_case ): if node is None: return MyNode(_snake_case ) if data < node.get_data(): node.set_left(insert_node(node.get_left() ,_snake_case ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected UpperCAmelCase__ : List[str] = node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child UpperCAmelCase__ : str = right_rotation(_snake_case ) else: UpperCAmelCase__ : Tuple = lr_rotation(_snake_case ) else: node.set_right(insert_node(node.get_right() ,_snake_case ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: UpperCAmelCase__ : Any = node.get_right() assert right_child is not None if data < right_child.get_data(): UpperCAmelCase__ : List[Any] = rl_rotation(_snake_case ) else: UpperCAmelCase__ : Tuple = left_rotation(_snake_case ) UpperCAmelCase__ : Union[str, Any] = my_max(get_height(node.get_right() ) ,get_height(node.get_left() ) ) + 1 node.set_height(_snake_case ) return node def lowerCamelCase ( _snake_case ): while True: UpperCAmelCase__ : List[Any] = root.get_right() if right_child is None: break UpperCAmelCase__ : Dict = right_child return root.get_data() def lowerCamelCase ( _snake_case ): while True: UpperCAmelCase__ : Union[str, Any] = root.get_left() if left_child is None: break UpperCAmelCase__ : List[Any] = left_child return root.get_data() def lowerCamelCase ( _snake_case ,_snake_case ): UpperCAmelCase__ : Tuple = root.get_left() UpperCAmelCase__ : List[str] = root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: UpperCAmelCase__ : Tuple = get_left_most(_snake_case ) root.set_data(_snake_case ) root.set_right(del_node(_snake_case ,_snake_case ) ) elif left_child is not None: UpperCAmelCase__ : Optional[int] = left_child elif right_child is not None: UpperCAmelCase__ : Tuple = right_child else: return None elif root.get_data() > data: if left_child is None: print('No such data' ) return root else: root.set_left(del_node(_snake_case ,_snake_case ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(_snake_case ,_snake_case ) ) if get_height(_snake_case ) - get_height(_snake_case ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): UpperCAmelCase__ : Dict = left_rotation(_snake_case ) else: UpperCAmelCase__ : Any = rl_rotation(_snake_case ) elif get_height(_snake_case ) - get_height(_snake_case ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): UpperCAmelCase__ : int = right_rotation(_snake_case ) else: UpperCAmelCase__ : int = lr_rotation(_snake_case ) UpperCAmelCase__ : Optional[Any] = my_max(get_height(root.get_right() ) ,get_height(root.get_left() ) ) + 1 root.set_height(_snake_case ) return root class a : def __init__( self ): UpperCAmelCase__ : MyNode | None = None def __snake_case ( self ): return get_height(self.root ) def __snake_case ( self , UpperCamelCase_ ): print('insert:' + str(UpperCamelCase_ ) ) UpperCAmelCase__ : Optional[int] = insert_node(self.root , UpperCamelCase_ ) def __snake_case ( self , UpperCamelCase_ ): print('delete:' + str(UpperCamelCase_ ) ) if self.root is None: print('Tree is empty!' ) return UpperCAmelCase__ : Any = del_node(self.root , UpperCamelCase_ ) def __str__( self , ): # a level traversale, gives a more intuitive look on the tree UpperCAmelCase__ : Optional[int] = '' UpperCAmelCase__ : Any = MyQueue() q.push(self.root ) UpperCAmelCase__ : Any = self.get_height() if layer == 0: return output UpperCAmelCase__ : Optional[Any] = 0 while not q.is_empty(): UpperCAmelCase__ : Any = q.pop() UpperCAmelCase__ : int = ' ' * int(math.pow(2 , layer - 1 ) ) output += space if node is None: output += "*" q.push(UpperCamelCase_ ) q.push(UpperCamelCase_ ) else: output += str(node.get_data() ) q.push(node.get_left() ) q.push(node.get_right() ) output += space UpperCAmelCase__ : Tuple = cnt + 1 for i in range(100 ): if cnt == math.pow(2 , UpperCamelCase_ ) - 1: UpperCAmelCase__ : str = layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def lowerCamelCase ( ): import doctest doctest.testmod() if __name__ == "__main__": _test() UpperCamelCase__ = AVLtree() UpperCamelCase__ = list(range(10)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
254
1
from __future__ import annotations from sys import maxsize from typing import Generic, TypeVar __snake_case :Union[str, Any] =TypeVar('T') def lowerCamelCase_ ( lowerCAmelCase__ : int ) -> int: '''simple docstring''' return (position - 1) // 2 def lowerCamelCase_ ( lowerCAmelCase__ : int ) -> int: '''simple docstring''' return (2 * position) + 1 def lowerCamelCase_ ( lowerCAmelCase__ : int ) -> int: '''simple docstring''' return (2 * position) + 2 class lowerCAmelCase__ ( Generic[T] ): def __init__( self : Any ) -> None: A = [] A = {} A = 0 def __len__( self : str ) -> int: return self.elements def __repr__( self : Dict ) -> str: return str(self.heap ) def __UpperCamelCase ( self : Tuple ) -> bool: # Check if the priority queue is empty return self.elements == 0 def __UpperCamelCase ( self : Optional[int] , __UpperCamelCase : T , __UpperCamelCase : int ) -> None: # Add an element with given priority to the queue self.heap.append((elem, weight) ) A = self.elements self.elements += 1 self._bubble_up(__UpperCamelCase ) def __UpperCamelCase ( self : Dict ) -> T: # Remove and return the element with lowest weight (highest priority) if self.elements > 1: self._swap_nodes(0 , self.elements - 1 ) A , A = self.heap.pop() del self.position_map[elem] self.elements -= 1 if self.elements > 0: A , A = self.heap[0] self._bubble_down(__UpperCamelCase ) return elem def __UpperCamelCase ( self : Any , __UpperCamelCase : T , __UpperCamelCase : int ) -> None: # Update the weight of the given key A = self.position_map[elem] A = (elem, weight) if position > 0: A = get_parent_position(__UpperCamelCase ) A , A = self.heap[parent_position] if parent_weight > weight: self._bubble_up(__UpperCamelCase ) else: self._bubble_down(__UpperCamelCase ) else: self._bubble_down(__UpperCamelCase ) def __UpperCamelCase ( self : Optional[int] , __UpperCamelCase : T ) -> None: # Place a node at the proper position (upward movement) [to be used internally # only] A = self.position_map[elem] if curr_pos == 0: return None A = get_parent_position(__UpperCamelCase ) A , A = self.heap[curr_pos] A , A = self.heap[parent_position] if parent_weight > weight: self._swap_nodes(__UpperCamelCase , __UpperCamelCase ) return self._bubble_up(__UpperCamelCase ) return None def __UpperCamelCase ( self : Tuple , __UpperCamelCase : T ) -> None: # Place a node at the proper position (downward movement) [to be used # internally only] A = self.position_map[elem] A , A = self.heap[curr_pos] A = get_child_left_position(__UpperCamelCase ) A = get_child_right_position(__UpperCamelCase ) if child_left_position < self.elements and child_right_position < self.elements: A , A = self.heap[child_left_position] A , A = self.heap[child_right_position] if child_right_weight < child_left_weight and child_right_weight < weight: self._swap_nodes(__UpperCamelCase , __UpperCamelCase ) return self._bubble_down(__UpperCamelCase ) if child_left_position < self.elements: A , A = self.heap[child_left_position] if child_left_weight < weight: self._swap_nodes(__UpperCamelCase , __UpperCamelCase ) return self._bubble_down(__UpperCamelCase ) else: return None if child_right_position < self.elements: A , A = self.heap[child_right_position] if child_right_weight < weight: self._swap_nodes(__UpperCamelCase , __UpperCamelCase ) return self._bubble_down(__UpperCamelCase ) return None def __UpperCamelCase ( self : List[str] , __UpperCamelCase : int , __UpperCamelCase : int ) -> None: # Swap the nodes at the given positions A = self.heap[nodea_pos][0] A = self.heap[nodea_pos][0] A , A = ( self.heap[nodea_pos], self.heap[nodea_pos], ) A = nodea_pos A = nodea_pos class lowerCAmelCase__ ( Generic[T] ): def __init__( self : List[str] ) -> None: A = {} A = 0 def __repr__( self : Optional[int] ) -> str: return str(self.connections ) def __len__( self : Union[str, Any] ) -> int: return self.nodes def __UpperCamelCase ( self : Optional[int] , __UpperCamelCase : T ) -> None: # Add a node in the graph if it is not in the graph if node not in self.connections: A = {} self.nodes += 1 def __UpperCamelCase ( self : Optional[int] , __UpperCamelCase : T , __UpperCamelCase : T , __UpperCamelCase : int ) -> None: # Add an edge between 2 nodes in the graph self.add_node(__UpperCamelCase ) self.add_node(__UpperCamelCase ) A = weight A = weight def lowerCamelCase_ ( lowerCAmelCase__ : GraphUndirectedWeighted[T] , ) -> tuple[dict[T, int], dict[T, T | None]]: '''simple docstring''' A = {node: maxsize for node in graph.connections} A = {node: None for node in graph.connections} A = MinPriorityQueue() for node, weight in dist.items(): priority_queue.push(lowerCAmelCase__ , lowerCAmelCase__ ) if priority_queue.is_empty(): return dist, parent # initialization A = priority_queue.extract_min() A = 0 for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: A = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(lowerCAmelCase__ , dist[neighbour] ) A = node # running prim's algorithm while not priority_queue.is_empty(): A = priority_queue.extract_min() for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: A = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(lowerCAmelCase__ , dist[neighbour] ) A = node return dist, parent
106
'''simple docstring''' from collections.abc import Callable def __snake_case (__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): """simple docstring""" lowerCamelCase_ : float = a lowerCamelCase_ : float = b if function(__UpperCAmelCase ) == 0: # one of the a or b is a root for the function return a elif function(__UpperCAmelCase ) == 0: return b elif ( function(__UpperCAmelCase ) * function(__UpperCAmelCase ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError('''could not find root in given interval.''' ) else: lowerCamelCase_ : float = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(__UpperCAmelCase ) == 0: return mid elif function(__UpperCAmelCase ) * function(__UpperCAmelCase ) < 0: lowerCamelCase_ : List[str] = mid else: lowerCamelCase_ : Any = mid lowerCamelCase_ : int = start + (end - start) / 2.0 return mid def __snake_case (__UpperCAmelCase ): """simple docstring""" return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 1000)) import doctest doctest.testmod()
501
0
from sklearn.metrics import fa_score import datasets __lowerCAmelCase : List[str] ='\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' __lowerCAmelCase : Dict ='\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`.\n\n - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {\'f1\': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results[\'f1\'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric("f1")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results[\'f1\'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro")\n >>> print(round(results[\'f1\'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted")\n >>> print(round(results[\'f1\'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {\'f1\': array([0.8, 0. , 0. ])}\n' __lowerCAmelCase : List[Any] ='\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): '''simple docstring''' def __magic_name__( self :Any ) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''int32''' ) ), '''references''': datasets.Sequence(datasets.Value('''int32''' ) ), } if self.config_name == '''multilabel''' else { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=['''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'''] , ) def __magic_name__( self :List[Any] , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Dict=None , lowerCAmelCase__ :Optional[Any]=1 , lowerCAmelCase__ :List[str]="binary" , lowerCAmelCase__ :str=None ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE : List[Any] = fa_score( UpperCAmelCase__ , UpperCAmelCase__ , labels=UpperCAmelCase__ , pos_label=UpperCAmelCase__ , average=UpperCAmelCase__ , sample_weight=UpperCAmelCase__ ) return {"f1": float(UpperCAmelCase__ ) if score.size == 1 else score}
718
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __lowerCAmelCase : Optional[Any] =argparse.ArgumentParser() parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument( '--txt2img_unclip', default='kakaobrain/karlo-v1-alpha', type=str, required=False, help='The pretrained txt2img unclip.', ) __lowerCAmelCase : str =parser.parse_args() __lowerCAmelCase : Tuple =UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __lowerCAmelCase : Optional[Any] =CLIPImageProcessor() __lowerCAmelCase : Union[str, Any] =CLIPVisionModelWithProjection.from_pretrained('openai/clip-vit-large-patch14') __lowerCAmelCase : Any =UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
260
0
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow lowerCAmelCase__ = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Path ,lowercase__ : Union[str, None] = None ,lowercase__ : Union[List[str], None] = None ,lowercase__ : Union[str, List[str], None] = None ,lowercase__ : bool = True ,): __lowercase = [file for file in os.listdir(lowercase__ ) if os.path.isfile(os.path.join(lowercase__ ,lowercase__ ) )] if identifier is not None: __lowercase = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(lowercase__ ,lowercase__ ): for n_ in n_identifier: __lowercase = [file for file in files if n_ not in file] else: __lowercase = [file for file in files if n_identifier not in file] __lowercase = ignore_files or [] ignore_files.append('''__init__.py''' ) __lowercase = [file for file in files if file not in ignore_files] for file in files: # Open all files print('''Testing''' ,lowercase__ ) if only_modules: __lowercase = file.split('''.''' )[0] try: __lowercase = getattr(lowercase__ ,lowercase__ ) __lowercase = doctest.DocTestSuite(lowercase__ ) __lowercase = unittest.TextTestRunner().run(lowercase__ ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"{module_identifier} is not a module." ) else: __lowercase = doctest.testfile(str('''..''' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = Path('''src/transformers''' ) __lowercase = '''modeling''' __lowercase = [ '''modeling_ctrl.py''', '''modeling_tf_ctrl.py''', ] self.analyze_directory(lowercase__ ,identifier=lowercase__ ,ignore_files=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = Path('''src/transformers''' ) __lowercase = '''tokenization''' self.analyze_directory(lowercase__ ,identifier=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = Path('''src/transformers''' ) __lowercase = '''configuration''' self.analyze_directory(lowercase__ ,identifier=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = Path('''src/transformers''' ) __lowercase = ['''configuration''', '''modeling''', '''tokenization'''] self.analyze_directory(lowercase__ ,n_identifier=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = Path('''docs/source''' ) __lowercase = ['''favicon.ico'''] self.analyze_directory(lowercase__ ,ignore_files=lowercase__ ,only_modules=lowercase__ )
41
'''simple docstring''' import json import sys def _lowerCAmelCase (_lowercase , _lowercase ): """simple docstring""" with open(_lowercase , encoding="utf-8" ) as f: a__ = json.load(_lowercase ) a__ = ["<details>", "<summary>Show updated benchmarks!</summary>", " "] for benchmark_name in sorted(_lowercase ): a__ = results[benchmark_name] a__ = benchmark_name.split("/" )[-1] output_md.append(F'### Benchmark: {benchmark_file_name}' ) a__ = "| metric |" a__ = "|--------|" a__ = "| new / old (diff) |" for metric_name in sorted(_lowercase ): a__ = benchmark_res[metric_name] a__ = metric_vals["new"] a__ = metric_vals.get("old" , _lowercase ) a__ = metric_vals.get("diff" , _lowercase ) a__ = F' {new_val:f}' if isinstance(_lowercase , (int, float) ) else "None" if old_val is not None: val_str += F' / {old_val:f}' if isinstance(_lowercase , (int, float) ) else "None" if dif_val is not None: val_str += F' ({dif_val:f})' if isinstance(_lowercase , (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append("</details>" ) with open(_lowercase , "w" , encoding="utf-8" ) as f: f.writelines("\n".join(_lowercase ) ) if __name__ == "__main__": UpperCamelCase_ : Dict = sys.argv[1] UpperCamelCase_ : int = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
331
0
'''simple docstring''' import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class _a ( __lowercase , unittest.TestCase ): """simple docstring""" A_ = CpmAntTokenizer A_ = False def lowerCamelCase__ ( self : str ): '''simple docstring''' super().setUp() lowercase_ = [ "<d>", "</d>", "<s>", "</s>", "</_>", "<unk>", "<pad>", "</n>", "我", "是", "C", "P", "M", "A", "n", "t", ] lowercase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) @tooslow def lowerCamelCase__ ( self : Tuple ): '''simple docstring''' lowercase_ = CpmAntTokenizer.from_pretrained("""openbmb/cpm-ant-10b""" ) lowercase_ = "今天天气真好!" lowercase_ = ["今天", "天气", "真", "好", "!"] lowercase_ = tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) lowercase_ = "今天天气真好!" lowercase_ = [tokenizer.bos_token] + tokens lowercase_ = [6, 9_802, 14_962, 2_082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase_ ) , lowercase_ ) lowercase_ = tokenizer.decode(lowercase_ ) self.assertEqual(lowercase_ , lowercase_ )
706
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy __snake_case = logging.get_logger(__name__) class _a ( __a ): """simple docstring""" def __init__( self : Dict , lowercase_ : int , lowercase_ : int , lowercase_ : float , **lowercase_ : Dict ): '''simple docstring''' lowercase_ = feature_size lowercase_ = sampling_rate lowercase_ = padding_value lowercase_ = kwargs.pop("""padding_side""" , """right""" ) lowercase_ = kwargs.pop("""return_attention_mask""" , lowercase_ ) super().__init__(**lowercase_ ) def lowerCamelCase__ ( self : List[Any] , lowercase_ : Union[ BatchFeature, List[BatchFeature], Dict[str, BatchFeature], Dict[str, List[BatchFeature]], List[Dict[str, BatchFeature]], ] , lowercase_ : Union[bool, str, PaddingStrategy] = True , lowercase_ : Optional[int] = None , lowercase_ : bool = False , lowercase_ : Optional[int] = None , lowercase_ : Optional[bool] = None , lowercase_ : Optional[Union[str, TensorType]] = None , ): '''simple docstring''' if isinstance(lowercase_ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): lowercase_ = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( """You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`""" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) lowercase_ = processed_features[self.model_input_names[0]] lowercase_ = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowercase_ ) == 0: if return_attention_mask: lowercase_ = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch lowercase_ = required_input[0] if isinstance(lowercase_ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. lowercase_ = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowercase_ ): lowercase_ = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowercase_ ): lowercase_ = """tf""" elif is_torch_tensor(lowercase_ ): lowercase_ = """pt""" elif isinstance(lowercase_ , (int, float, list, tuple, np.ndarray) ): lowercase_ = """np""" else: raise ValueError( F"""type of {first_element} unknown: {type(lowercase_ )}. """ """Should be one of a python, numpy, pytorch or tensorflow object.""" ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): lowercase_ = to_numpy(lowercase_ ) else: lowercase_ = [to_numpy(lowercase_ ) for v in value] # Convert padding_strategy in PaddingStrategy lowercase_ = self._get_padding_strategies(padding=lowercase_ , max_length=lowercase_ ) lowercase_ = processed_features[self.model_input_names[0]] lowercase_ = len(lowercase_ ) if not all(len(lowercase_ ) == batch_size for v in processed_features.values() ): raise ValueError("""Some items in the output dictionary have a different batch size than others.""" ) lowercase_ = [] for i in range(lowercase_ ): lowercase_ = {k: v[i] for k, v in processed_features.items()} # truncation lowercase_ = self._truncate( lowercase_ , max_length=lowercase_ , pad_to_multiple_of=lowercase_ , truncation=lowercase_ , ) truncated_inputs.append(lowercase_ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length lowercase_ = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) lowercase_ = PaddingStrategy.MAX_LENGTH lowercase_ = {} for i in range(lowercase_ ): # padding lowercase_ = self._pad( truncated_inputs[i] , max_length=lowercase_ , padding_strategy=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , ) for key, value in outputs.items(): if key not in batch_outputs: lowercase_ = [] if value.dtype is np.dtype(np.floataa ): lowercase_ = value.astype(np.floataa ) batch_outputs[key].append(lowercase_ ) return BatchFeature(lowercase_ , tensor_type=lowercase_ ) def lowerCamelCase__ ( self : Any , lowercase_ : Union[Dict[str, np.ndarray], BatchFeature] , lowercase_ : Optional[int] = None , lowercase_ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , lowercase_ : Optional[int] = None , lowercase_ : Optional[bool] = None , ): '''simple docstring''' lowercase_ = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: lowercase_ = len(lowercase_ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): lowercase_ = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of lowercase_ = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowercase_ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: lowercase_ = np.ones(len(lowercase_ ) , dtype=np.intaa ) if needs_to_be_padded: lowercase_ = max_length - len(lowercase_ ) if self.padding_side == "right": if return_attention_mask: lowercase_ = np.pad( processed_features["""attention_mask"""] , (0, difference) ) lowercase_ = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) lowercase_ = np.pad( lowercase_ , lowercase_ , """constant""" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: lowercase_ = np.pad( processed_features["""attention_mask"""] , (difference, 0) ) lowercase_ = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) lowercase_ = np.pad( lowercase_ , lowercase_ , """constant""" , constant_values=self.padding_value ) else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return processed_features def lowerCamelCase__ ( self : List[Any] , lowercase_ : Union[Dict[str, np.ndarray], BatchFeature] , lowercase_ : Optional[int] = None , lowercase_ : Optional[int] = None , lowercase_ : Optional[bool] = None , ): '''simple docstring''' if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("""When setting ``truncation=True``, make sure that ``max_length`` is defined.""" ) lowercase_ = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): lowercase_ = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of lowercase_ = len(lowercase_ ) > max_length if needs_to_be_truncated: lowercase_ = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: lowercase_ = processed_features["""attention_mask"""][:max_length] return processed_features def lowerCamelCase__ ( self : List[Any] , lowercase_ : Optional[int]=False , lowercase_ : List[str]=None ): '''simple docstring''' if padding is not False: if padding is True: lowercase_ = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowercase_ , lowercase_ ): lowercase_ = PaddingStrategy(lowercase_ ) elif isinstance(lowercase_ , lowercase_ ): lowercase_ = padding else: lowercase_ = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( """Asking to pad but the feature_extractor does not have a padding value. Please select a value to use""" """ as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.""" ) return padding_strategy
603
0
import math def lowerCAmelCase__(__snake_case ) -> bool: '''simple docstring''' return math.sqrt(__snake_case ) * math.sqrt(__snake_case ) == num def lowerCAmelCase__(__snake_case ) -> bool: '''simple docstring''' lowerCamelCase__ = 0 lowerCamelCase__ = n while left <= right: lowerCamelCase__ = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: lowerCamelCase__ = mid - 1 else: lowerCamelCase__ = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
481
import logging from transformers import PretrainedConfig _a = logging.getLogger(__name__) _a = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class __A ( lowerCAmelCase ): '''simple docstring''' lowerCAmelCase_ = """bertabs""" def __init__( self , __lowerCAmelCase=3_0_5_2_2 , __lowerCAmelCase=5_1_2 , __lowerCAmelCase=6 , __lowerCAmelCase=5_1_2 , __lowerCAmelCase=8 , __lowerCAmelCase=5_1_2 , __lowerCAmelCase=0.2 , __lowerCAmelCase=6 , __lowerCAmelCase=7_6_8 , __lowerCAmelCase=8 , __lowerCAmelCase=2_0_4_8 , __lowerCAmelCase=0.2 , **__lowerCAmelCase , ): '''simple docstring''' super().__init__(**__lowerCAmelCase ) lowerCamelCase__ = vocab_size lowerCamelCase__ = max_pos lowerCamelCase__ = enc_layers lowerCamelCase__ = enc_hidden_size lowerCamelCase__ = enc_heads lowerCamelCase__ = enc_ff_size lowerCamelCase__ = enc_dropout lowerCamelCase__ = dec_layers lowerCamelCase__ = dec_hidden_size lowerCamelCase__ = dec_heads lowerCamelCase__ = dec_ff_size lowerCamelCase__ = dec_dropout
481
1
def UpperCAmelCase__ ( lowercase__ ) -> list[int]: if length <= 0 or not isinstance(lowercase__ , lowercase__ ): raise ValueError("""Length must be a positive integer.""" ) return [n * (2 * n - 1) for n in range(lowercase__ )] if __name__ == "__main__": print(hexagonal_numbers(length=5)) print(hexagonal_numbers(length=10))
634
from __future__ import annotations from collections.abc import Callable UpperCamelCase__ = list[list[float | int]] def UpperCAmelCase__ ( lowercase__ , lowercase__ ) -> Matrix: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): __lowercase = matrix[row][col] __lowercase = vector[row][0] __lowercase = 0 __lowercase = 0 while row < size and col < size: # pivoting __lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: __lowercase , __lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): __lowercase = augmented[rowa][col] / augmented[row][col] __lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): __lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def UpperCAmelCase__ ( lowercase__ ) -> Callable[[int], int]: __lowercase = len(lowercase__ ) __lowercase = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] __lowercase = [[0] for _ in range(lowercase__ )] __lowercase = 42 __lowercase = 42 __lowercase = 42 __lowercase = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): __lowercase = (x_val + 1) ** (size - col - 1) __lowercase = y_val __lowercase = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowercase__ ) -> int: return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def UpperCAmelCase__ ( lowercase__ = question_function , lowercase__ = 10 ) -> int: __lowercase = [func(lowercase__ ) for x_val in range(1 , order + 1 )] __lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] __lowercase = 0 __lowercase = 42 __lowercase = 42 for poly in polynomials: __lowercase = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
634
1
def lowercase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: SCREAMING_SNAKE_CASE_ = _modexpt(SCREAMING_SNAKE_CASE , exponent // 2 , SCREAMING_SNAKE_CASE ) % modulo_value return (x * x) % modulo_value else: return (base * _modexpt(SCREAMING_SNAKE_CASE , exponent - 1 , SCREAMING_SNAKE_CASE )) % modulo_value def lowercase ( SCREAMING_SNAKE_CASE = 17_77 , SCREAMING_SNAKE_CASE = 18_55 , SCREAMING_SNAKE_CASE = 8 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE_ = base for _ in range(1 , SCREAMING_SNAKE_CASE ): SCREAMING_SNAKE_CASE_ = _modexpt(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , 10**digits ) return result if __name__ == "__main__": print(f"""{solution() = }""")
205
def lowercase ( SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_ = len(SCREAMING_SNAKE_CASE ) while cur > 1: # Find the maximum number in arr SCREAMING_SNAKE_CASE_ = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi SCREAMING_SNAKE_CASE_ = arr[mi::-1] + arr[mi + 1 : len(SCREAMING_SNAKE_CASE )] # Reverse whole list SCREAMING_SNAKE_CASE_ = arr[cur - 1 :: -1] + arr[cur : len(SCREAMING_SNAKE_CASE )] cur -= 1 return arr if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : List[str] = input("Enter numbers separated by a comma:\n").strip() SCREAMING_SNAKE_CASE__ : Optional[int] = [int(item) for item in user_input.split(",")] print(pancake_sort(unsorted))
205
1
import math def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(SCREAMING_SNAKE_CASE_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("This should never happen" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. lowercase_ = """Enter the base and the power separated by a comma: """ lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) lowercase_ , lowercase_ = map(int, input(prompt).split(""",""")) # We find the log of each number, using the function res(), which takes two # arguments. lowercase_ = res(xa, ya) lowercase_ = res(xa, ya) # We check for the largest number if resa > resa: print("""Largest number is""", xa, """^""", ya) elif resa > resa: print("""Largest number is""", xa, """^""", ya) else: print("""Both are equal""")
37
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ , lowercase__ = len(SCREAMING_SNAKE_CASE_ ), len(grid[0] ) if ( min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) lowercase__ = 0 count += depth_first_search(SCREAMING_SNAKE_CASE_ , row + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , row - 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col + 1 , SCREAMING_SNAKE_CASE_ ) count += depth_first_search(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , col - 1 , SCREAMING_SNAKE_CASE_ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
37
1