code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations import unittest from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel @require_tf class __lowercase : """simple docstring""" _UpperCAmelCase = BlenderbotSmallConfig _UpperCAmelCase = {} _UpperCAmelCase = """gelu""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=1_3 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=9_9 , lowerCAmelCase__=3_2 , lowerCAmelCase__=2 , lowerCAmelCase__=4 , lowerCAmelCase__=3_7 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=2_0 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=0 , ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = parent SCREAMING_SNAKE_CASE_ : List[str] = batch_size SCREAMING_SNAKE_CASE_ : int = seq_length SCREAMING_SNAKE_CASE_ : Any = is_training SCREAMING_SNAKE_CASE_ : List[Any] = use_labels SCREAMING_SNAKE_CASE_ : List[str] = vocab_size SCREAMING_SNAKE_CASE_ : Tuple = hidden_size SCREAMING_SNAKE_CASE_ : int = num_hidden_layers SCREAMING_SNAKE_CASE_ : Any = num_attention_heads SCREAMING_SNAKE_CASE_ : List[Any] = intermediate_size SCREAMING_SNAKE_CASE_ : Optional[Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : Optional[Any] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : int = max_position_embeddings SCREAMING_SNAKE_CASE_ : Any = eos_token_id SCREAMING_SNAKE_CASE_ : int = pad_token_id SCREAMING_SNAKE_CASE_ : Dict = bos_token_id def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) SCREAMING_SNAKE_CASE_ : Tuple = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) SCREAMING_SNAKE_CASE_ : str = tf.concat([input_ids, eos_tensor] , axis=1 ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE_ : str = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) SCREAMING_SNAKE_CASE_ : List[Any] = prepare_blenderbot_small_inputs_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return config, inputs_dict def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = TFBlenderbotSmallModel(config=lowerCAmelCase__ ).get_decoder() SCREAMING_SNAKE_CASE_ : Optional[Any] = inputs_dict['input_ids'] SCREAMING_SNAKE_CASE_ : Dict = input_ids[:1, :] SCREAMING_SNAKE_CASE_ : str = inputs_dict['attention_mask'][:1, :] SCREAMING_SNAKE_CASE_ : str = inputs_dict['head_mask'] SCREAMING_SNAKE_CASE_ : List[str] = 1 # first forward pass SCREAMING_SNAKE_CASE_ : Tuple = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , head_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : List[Any] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids SCREAMING_SNAKE_CASE_ : List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE_ : List[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and SCREAMING_SNAKE_CASE_ : str = tf.concat([input_ids, next_tokens] , axis=-1 ) SCREAMING_SNAKE_CASE_ : Any = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) SCREAMING_SNAKE_CASE_ : str = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ )[0] SCREAMING_SNAKE_CASE_ : List[str] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice SCREAMING_SNAKE_CASE_ : Optional[int] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) SCREAMING_SNAKE_CASE_ : Optional[int] = output_from_no_past[:, -3:, random_slice_idx] SCREAMING_SNAKE_CASE_ : Optional[int] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase__ , lowerCAmelCase__ , rtol=1E-3 ) def a__ ( A__, A__, A__, A__=None, A__=None, A__=None, A__=None, A__=None, ): if attention_mask is None: SCREAMING_SNAKE_CASE_ : str = tf.cast(tf.math.not_equal(A__, config.pad_token_id ), tf.inta ) if decoder_attention_mask is None: SCREAMING_SNAKE_CASE_ : List[Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape, dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:], config.pad_token_id ), tf.inta ), ], axis=-1, ) if head_mask is None: SCREAMING_SNAKE_CASE_ : Optional[Any] = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: SCREAMING_SNAKE_CASE_ : Dict = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: SCREAMING_SNAKE_CASE_ : List[str] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __lowercase (__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" _UpperCAmelCase = ( (TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel) if is_tf_available() else () ) _UpperCAmelCase = (TFBlenderbotSmallForConditionalGeneration,) if is_tf_available() else () _UpperCAmelCase = ( { """conversational""": TFBlenderbotSmallForConditionalGeneration, """feature-extraction""": TFBlenderbotSmallModel, """summarization""": TFBlenderbotSmallForConditionalGeneration, """text2text-generation""": TFBlenderbotSmallForConditionalGeneration, """translation""": TFBlenderbotSmallForConditionalGeneration, } if is_tf_available() else {} ) _UpperCAmelCase = True _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = TFBlenderbotSmallModelTester(self ) SCREAMING_SNAKE_CASE_ : List[str] = ConfigTester(self , config_class=lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase__ ) @require_tokenizers @require_tf class __lowercase (unittest.TestCase ): """simple docstring""" _UpperCAmelCase = [ """Social anxiety\nWow, I am never shy. Do you have anxiety?\nYes. I end up sweating and blushing and feel like """ """ i'm going to throw up.\nand why is that?""" ] _UpperCAmelCase = """facebook/blenderbot_small-90M""" @cached_property def UpperCamelCase__ ( self ): """simple docstring""" return BlenderbotSmallTokenizer.from_pretrained('facebook/blenderbot-90M' ) @cached_property def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.tokenizer(self.src_text , return_tensors='tf' ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=lowerCAmelCase__ , ) SCREAMING_SNAKE_CASE_ : Any = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=lowerCAmelCase__ )[0] assert generated_words in ( "i don't know. i just feel like i'm going to throw up. it's not fun.", "i'm not sure. i just feel like i've been feeling like i have to be in a certain place", "i'm not sure. i just feel like i've been in a bad situation.", )
101
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
0
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __magic_name__ : Tuple = { """configuration_efficientnet""": [ """EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientNetConfig""", """EfficientNetOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : Optional[int] = ["""EfficientNetImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ : List[str] = [ """EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientNetForImageClassification""", """EfficientNetModel""", """EfficientNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys __magic_name__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
102
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
0
"""simple docstring""" from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
103
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
"""simple docstring""" import inspect from typing import Callable, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import DiffusionPipeline from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import logging UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" def __init__( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) -> Union[str, Any]: super().__init__() self.register_modules( vae=SCREAMING_SNAKE_CASE__ , text_encoder=SCREAMING_SNAKE_CASE__ , tokenizer=SCREAMING_SNAKE_CASE__ , unet=SCREAMING_SNAKE_CASE__ , scheduler=SCREAMING_SNAKE_CASE__ , safety_checker=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ , ) def snake_case__ ( self , SCREAMING_SNAKE_CASE__ = "auto" ) -> int: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory A__ = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(SCREAMING_SNAKE_CASE__ ) def snake_case__ ( self ) -> Union[str, Any]: self.enable_attention_slicing(SCREAMING_SNAKE_CASE__ ) @torch.no_grad() def __call__( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 512 , SCREAMING_SNAKE_CASE__ = 512 , SCREAMING_SNAKE_CASE__ = 50 , SCREAMING_SNAKE_CASE__ = 7.5 , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = 0.0 , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = "pil" , SCREAMING_SNAKE_CASE__ = True , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = None , **SCREAMING_SNAKE_CASE__ , ) -> Dict: if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): A__ = 1 elif isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): A__ = len(SCREAMING_SNAKE_CASE__ ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(SCREAMING_SNAKE_CASE__ )}""" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f"""`height` and `width` have to be divisible by 8 but are {height} and {width}.""" ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(SCREAMING_SNAKE_CASE__ )}.""" ) # get prompt text embeddings A__ = self.tokenizer( SCREAMING_SNAKE_CASE__ , padding="max_length" , max_length=self.tokenizer.model_max_length , return_tensors="pt" , ) A__ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: A__ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( "The following part of your input was truncated because CLIP can only handle sequences up to" f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) A__ = text_input_ids[:, : self.tokenizer.model_max_length] if text_embeddings is None: A__ = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method A__ , A__ , A__ = text_embeddings.shape A__ = text_embeddings.repeat(1 , SCREAMING_SNAKE_CASE__ , 1 ) A__ = text_embeddings.view(bs_embed * num_images_per_prompt , SCREAMING_SNAKE_CASE__ , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. A__ = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: A__ = 42 if negative_prompt is None: A__ = [""] elif type(SCREAMING_SNAKE_CASE__ ) is not type(SCREAMING_SNAKE_CASE__ ): raise TypeError( f"""`negative_prompt` should be the same type to `prompt`, but got {type(SCREAMING_SNAKE_CASE__ )} !=""" f""" {type(SCREAMING_SNAKE_CASE__ )}.""" ) elif isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): A__ = [negative_prompt] elif batch_size != len(SCREAMING_SNAKE_CASE__ ): raise ValueError( f"""`negative_prompt`: {negative_prompt} has batch size {len(SCREAMING_SNAKE_CASE__ )}, but `prompt`:""" f""" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches""" " the batch size of `prompt`." ) else: A__ = negative_prompt A__ = text_input_ids.shape[-1] A__ = self.tokenizer( SCREAMING_SNAKE_CASE__ , padding="max_length" , max_length=SCREAMING_SNAKE_CASE__ , truncation=SCREAMING_SNAKE_CASE__ , return_tensors="pt" , ) A__ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method A__ = uncond_embeddings.shape[1] A__ = uncond_embeddings.repeat(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 ) A__ = uncond_embeddings.view(batch_size * num_images_per_prompt , SCREAMING_SNAKE_CASE__ , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes A__ = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. A__ = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) A__ = (batch_size * num_images_per_prompt, self.unet.config.in_channels, 64, 64) A__ = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps A__ = torch.randn( SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , device="cpu" , dtype=SCREAMING_SNAKE_CASE__ ).to(self.device ) A__ = torch.randn(SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , device="cpu" , dtype=SCREAMING_SNAKE_CASE__ ).to( self.device ) else: A__ = torch.randn( SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , device=self.device , dtype=SCREAMING_SNAKE_CASE__ ) A__ = torch.randn(SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , device=self.device , dtype=SCREAMING_SNAKE_CASE__ ) else: if latents_reference.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) A__ = latents_reference.to(self.device ) A__ = latents.to(self.device ) # This is the key part of the pipeline where we # try to ensure that the generated images w/ the same seed # but different sizes actually result in similar images A__ = (latents_shape[3] - latents_shape_reference[3]) // 2 A__ = (latents_shape[2] - latents_shape_reference[2]) // 2 A__ = latents_shape_reference[3] if dx >= 0 else latents_shape_reference[3] + 2 * dx A__ = latents_shape_reference[2] if dy >= 0 else latents_shape_reference[2] + 2 * dy A__ = 0 if dx < 0 else dx A__ = 0 if dy < 0 else dy A__ = max(-dx , 0 ) A__ = max(-dy , 0 ) # import pdb # pdb.set_trace() A__ = latents_reference[:, :, dy : dy + h, dx : dx + w] # set timesteps self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE__ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand A__ = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler A__ = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] A__ = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) A__ = {} if accepts_eta: A__ = eta for i, t in enumerate(self.progress_bar(SCREAMING_SNAKE_CASE__ ) ): # expand the latents if we are doing classifier free guidance A__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents A__ = self.scheduler.scale_model_input(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # predict the noise residual A__ = self.unet(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , encoder_hidden_states=SCREAMING_SNAKE_CASE__ ).sample # perform guidance if do_classifier_free_guidance: A__ , A__ = noise_pred.chunk(2 ) A__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 A__ = self.scheduler.step(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) A__ = 1 / 0.1_8_2_1_5 * latents A__ = self.vae.decode(SCREAMING_SNAKE_CASE__ ).sample A__ = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 A__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if self.safety_checker is not None: A__ = self.feature_extractor(self.numpy_to_pil(SCREAMING_SNAKE_CASE__ ) , return_tensors="pt" ).to( self.device ) A__ , A__ = self.safety_checker( images=SCREAMING_SNAKE_CASE__ , clip_input=safety_checker_input.pixel_values.to(text_embeddings.dtype ) ) else: A__ = None if output_type == "pil": A__ = self.numpy_to_pil(SCREAMING_SNAKE_CASE__ ) if not return_dict: return (image, has_nsfw_concept) return StableDiffusionPipelineOutput(images=SCREAMING_SNAKE_CASE__ , nsfw_content_detected=SCREAMING_SNAKE_CASE__ )
104
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore UpperCamelCase__ : Union[str, Any] = ''' Human: <<task>> Assistant: ''' UpperCamelCase__ : Optional[int] = '''huggingface-tools/default-prompts''' UpperCamelCase__ : str = {'''chat''': '''chat_prompt_template.txt''', '''run''': '''run_prompt_template.txt'''} def __UpperCAmelCase ( lowerCamelCase_ : Tuple , lowerCamelCase_ : str , lowerCamelCase_ : List[str]="run" ) -> List[Any]: """simple docstring""" if prompt_or_repo_id is None: SCREAMING_SNAKE_CASE_ : int = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search('\\s' , lowerCamelCase_ ) is not None: return prompt_or_repo_id SCREAMING_SNAKE_CASE_ : List[Any] = cached_file( lowerCamelCase_ , PROMPT_FILES[mode] , repo_type='dataset' , user_agent={'agent': agent_name} ) with open(lowerCamelCase_ , 'r' , encoding='utf-8' ) as f: return f.read()
105
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
0
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __snake_case :Any =logging.get_logger(__name__) __snake_case :str ={ 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class lowerCAmelCase__ ( _lowerCamelCase ): def __init__( self : Tuple , __UpperCamelCase : Any=None , __UpperCamelCase : List[Any]=None , *__UpperCamelCase : Union[str, Any] , **__UpperCamelCase : Optional[Any] ) -> Tuple: super().__init__(*__UpperCamelCase , **__UpperCamelCase ) if config is None: assert isinstance(self.model , __UpperCamelCase ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" f''' {self.model.__class__}''' ) A = self.model.config else: A = config A = data_args A = self.config.tgt_vocab_size if isinstance(self.config , __UpperCamelCase ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( f'''The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for''' ' padding..' ) if self.args.label_smoothing == 0: A = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss A = label_smoothed_nll_loss def __UpperCamelCase ( self : int , __UpperCamelCase : int ) -> str: if self.optimizer is None: A = ['bias', 'LayerNorm.weight'] A = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] A = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: A = Adafactor A = {'scale_parameter': False, 'relative_step': False} else: A = AdamW A = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } A = self.args.learning_rate if self.sharded_ddp: A = OSS( params=__UpperCamelCase , optim=__UpperCamelCase , **__UpperCamelCase , ) else: A = optimizer_cls(__UpperCamelCase , **__UpperCamelCase ) if self.lr_scheduler is None: A = self._get_lr_scheduler(__UpperCamelCase ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __UpperCamelCase ( self : str , __UpperCamelCase : List[Any] ) -> Union[str, Any]: A = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": A = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": A = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: A = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=__UpperCamelCase ) return scheduler def __UpperCamelCase ( self : List[Any] ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __UpperCamelCase ( self : Optional[int] , __UpperCamelCase : List[str] , __UpperCamelCase : Dict , __UpperCamelCase : Optional[int] ) -> int: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token A = model(**__UpperCamelCase , use_cache=__UpperCamelCase )[0] A = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models A , A = model(**__UpperCamelCase , labels=__UpperCamelCase , use_cache=__UpperCamelCase )[:2] else: # compute label smoothed loss A = model(**__UpperCamelCase , use_cache=__UpperCamelCase )[0] A = torch.nn.functional.log_softmax(__UpperCamelCase , dim=-1 ) A , A = self.loss_fn(__UpperCamelCase , __UpperCamelCase , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def __UpperCamelCase ( self : str , __UpperCamelCase : Tuple , __UpperCamelCase : List[Any] ) -> List[Any]: A = inputs.pop('labels' ) A , A = self._compute_loss(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return loss def __UpperCamelCase ( self : int , __UpperCamelCase : nn.Module , __UpperCamelCase : Dict[str, Union[torch.Tensor, Any]] , __UpperCamelCase : bool , __UpperCamelCase : Optional[List[str]] = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: A = self._prepare_inputs(__UpperCamelCase ) A = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: A = self.model.generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , **__UpperCamelCase , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: A = self._pad_tensors_to_max_len(__UpperCamelCase , gen_kwargs['max_length'] ) A = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data A , A = self._compute_loss(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) A = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) A = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: A = self._pad_tensors_to_max_len(__UpperCamelCase , gen_kwargs['max_length'] ) return (loss, logits, labels) def __UpperCamelCase ( self : Dict , __UpperCamelCase : List[str] , __UpperCamelCase : Dict ) -> int: # If PAD token is not defined at least EOS token has to be defined A = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' f''' padded to `max_length`={max_length}''' ) A = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) A = tensor return padded_tensor
106
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
0
'''simple docstring''' import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ : """simple docstring""" def __init__( self : List[str], UpperCamelCase__ : List[str], UpperCamelCase__ : List[str]=13, UpperCamelCase__ : str=[30, 30], UpperCamelCase__ : List[Any]=2, UpperCamelCase__ : Optional[int]=3, UpperCamelCase__ : Dict=True, UpperCamelCase__ : int=True, UpperCamelCase__ : List[str]=32, UpperCamelCase__ : List[Any]=5, UpperCamelCase__ : Tuple=4, UpperCamelCase__ : Optional[Any]=37, UpperCamelCase__ : Tuple="gelu", UpperCamelCase__ : List[Any]=0.1, UpperCamelCase__ : List[Any]=0.1, UpperCamelCase__ : Dict=10, UpperCamelCase__ : List[Any]=0.02, UpperCamelCase__ : Optional[int]=3, UpperCamelCase__ : str=None, UpperCamelCase__ : List[str]=8, UpperCamelCase__ : Tuple=10, ) -> Any: _A = parent _A = batch_size _A = image_size _A = patch_size _A = num_channels _A = is_training _A = use_labels _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = type_sequence_label_size _A = initializer_range _A = num_labels _A = scope _A = n_targets _A = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens _A = (image_size[1] // patch_size) * (image_size[0] // patch_size) _A = num_patches + 1 + self.num_detection_tokens def __UpperCAmelCase ( self : str ) -> str: _A = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) _A = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) _A = [] for i in range(self.batch_size ): _A = {} _A = torch.randint( high=self.num_labels, size=(self.n_targets,), device=UpperCamelCase__ ) _A = torch.rand(self.n_targets, 4, device=UpperCamelCase__ ) labels.append(UpperCamelCase__ ) _A = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self : int ) -> Any: return YolosConfig( image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=UpperCamelCase__, initializer_range=self.initializer_range, num_detection_tokens=self.num_detection_tokens, num_labels=self.num_labels, ) def __UpperCAmelCase ( self : Union[str, Any], UpperCamelCase__ : str, UpperCamelCase__ : Optional[int], UpperCamelCase__ : Optional[int] ) -> Optional[Any]: _A = YolosModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() _A = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.expected_seq_len, self.hidden_size) ) def __UpperCAmelCase ( self : Tuple, UpperCamelCase__ : str, UpperCamelCase__ : List[Any], UpperCamelCase__ : Optional[Any] ) -> Any: _A = YolosForObjectDetection(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() _A = model(pixel_values=UpperCamelCase__ ) _A = model(UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape, (self.batch_size, self.num_detection_tokens, 4) ) _A = model(pixel_values=UpperCamelCase__, labels=UpperCamelCase__ ) self.parent.assertEqual(result.loss.shape, () ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape, (self.batch_size, self.num_detection_tokens, 4) ) def __UpperCAmelCase ( self : Tuple ) -> Union[str, Any]: _A = self.prepare_config_and_inputs() _A , _A , _A = config_and_inputs _A = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowercase_ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): """simple docstring""" __lowerCAmelCase = (YolosModel, YolosForObjectDetection) if is_torch_available() else () __lowerCAmelCase = ( {"feature-extraction": YolosModel, "object-detection": YolosForObjectDetection} if is_torch_available() else {} ) __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def __UpperCAmelCase ( self : str, UpperCamelCase__ : List[Any], UpperCamelCase__ : Dict, UpperCamelCase__ : List[Any]=False ) -> Union[str, Any]: _A = super()._prepare_for_class(UpperCamelCase__, UpperCamelCase__, return_labels=UpperCamelCase__ ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": _A = [] for i in range(self.model_tester.batch_size ): _A = {} _A = torch.ones( size=(self.model_tester.n_targets,), device=UpperCamelCase__, dtype=torch.long ) _A = torch.ones( self.model_tester.n_targets, 4, device=UpperCamelCase__, dtype=torch.float ) labels.append(UpperCamelCase__ ) _A = labels return inputs_dict def __UpperCAmelCase ( self : Any ) -> Dict: _A = YolosModelTester(self ) _A = ConfigTester(self, config_class=UpperCamelCase__, has_text_modality=UpperCamelCase__, hidden_size=37 ) def __UpperCAmelCase ( self : Optional[int] ) -> Optional[int]: self.config_tester.run_common_tests() def __UpperCAmelCase ( self : str ) -> int: # YOLOS does not use inputs_embeds pass def __UpperCAmelCase ( self : str ) -> str: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(UpperCamelCase__ ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) _A = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase__, nn.Linear ) ) def __UpperCAmelCase ( self : List[str] ) -> int: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = model_class(UpperCamelCase__ ) _A = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _A = [*signature.parameters.keys()] _A = ['pixel_values'] self.assertListEqual(arg_names[:1], UpperCamelCase__ ) def __UpperCAmelCase ( self : Optional[int] ) -> Any: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def __UpperCAmelCase ( self : Optional[Any] ) -> List[str]: _A , _A = self.model_tester.prepare_config_and_inputs_for_common() _A = True # in YOLOS, the seq_len is different _A = self.model_tester.expected_seq_len for model_class in self.all_model_classes: _A = True _A = False _A = True _A = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _A = model(**self._prepare_for_class(UpperCamelCase__, UpperCamelCase__ ) ) _A = outputs.attentions self.assertEqual(len(UpperCamelCase__ ), self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _A = True _A = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _A = model(**self._prepare_for_class(UpperCamelCase__, UpperCamelCase__ ) ) _A = outputs.attentions self.assertEqual(len(UpperCamelCase__ ), self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ), [self.model_tester.num_attention_heads, seq_len, seq_len], ) _A = len(UpperCamelCase__ ) # Check attention is always last and order is fine _A = True _A = True _A = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _A = model(**self._prepare_for_class(UpperCamelCase__, UpperCamelCase__ ) ) _A = 1 self.assertEqual(out_len + added_hidden_states, len(UpperCamelCase__ ) ) _A = outputs.attentions self.assertEqual(len(UpperCamelCase__ ), self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ), [self.model_tester.num_attention_heads, seq_len, seq_len], ) def __UpperCAmelCase ( self : Dict ) -> Dict: def check_hidden_states_output(UpperCamelCase__ : int, UpperCamelCase__ : Tuple, UpperCamelCase__ : str ): _A = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): _A = model(**self._prepare_for_class(UpperCamelCase__, UpperCamelCase__ ) ) _A = outputs.hidden_states _A = getattr( self.model_tester, 'expected_num_hidden_layers', self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(UpperCamelCase__ ), UpperCamelCase__ ) # YOLOS has a different seq_length _A = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ), [seq_length, self.model_tester.hidden_size], ) _A , _A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A = True check_hidden_states_output(UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _A = True check_hidden_states_output(UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ ) def __UpperCAmelCase ( self : str ) -> Optional[Any]: _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*UpperCamelCase__ ) @slow def __UpperCAmelCase ( self : Optional[int] ) -> str: for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A = YolosModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def _SCREAMING_SNAKE_CASE ( ): _A = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowercase_ ( unittest.TestCase ): """simple docstring""" @cached_property def __UpperCAmelCase ( self : Dict ) -> Any: return AutoImageProcessor.from_pretrained('hustvl/yolos-small' ) if is_vision_available() else None @slow def __UpperCAmelCase ( self : List[Any] ) -> int: _A = YolosForObjectDetection.from_pretrained('hustvl/yolos-small' ).to(UpperCamelCase__ ) _A = self.default_image_processor _A = prepare_img() _A = image_processor(images=UpperCamelCase__, return_tensors='pt' ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): _A = model(inputs.pixel_values ) # verify outputs _A = torch.Size((1, 1_00, 92) ) self.assertEqual(outputs.logits.shape, UpperCamelCase__ ) _A = torch.tensor( [[-24.0_248, -10.3_024, -14.8_290], [-42.0_392, -16.8_200, -27.4_334], [-27.2_743, -11.8_154, -18.7_148]], device=UpperCamelCase__, ) _A = torch.tensor( [[0.2_559, 0.5_455, 0.4_706], [0.2_989, 0.7_279, 0.1_875], [0.7_732, 0.4_017, 0.4_462]], device=UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3], UpperCamelCase__, atol=1e-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3], UpperCamelCase__, atol=1e-4 ) ) # verify postprocessing _A = image_processor.post_process_object_detection( UpperCamelCase__, threshold=0.3, target_sizes=[image.size[::-1]] )[0] _A = torch.tensor([0.9_994, 0.9_790, 0.9_964, 0.9_972, 0.9_861] ).to(UpperCamelCase__ ) _A = [75, 75, 17, 63, 17] _A = torch.tensor([335.0_609, 79.3_848, 375.4_216, 187.2_495] ).to(UpperCamelCase__ ) self.assertEqual(len(results['scores'] ), 5 ) self.assertTrue(torch.allclose(results['scores'], UpperCamelCase__, atol=1e-4 ) ) self.assertSequenceEqual(results['labels'].tolist(), UpperCamelCase__ ) self.assertTrue(torch.allclose(results['boxes'][0, :], UpperCamelCase__ ) )
107
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
0
from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def lowerCamelCase ( self : List[str] ) -> Dict: """simple docstring""" return [ {"col_1": 3, "col_2": "a"}, {"col_1": 2, "col_2": "b"}, {"col_1": 1, "col_2": "c"}, {"col_1": 0, "col_2": "d"}, ] def lowerCamelCase ( self : List[str] ) -> List[Any]: """simple docstring""" _UpperCAmelCase = {"""col_1""": [3, 2, 1, 0], """col_2""": ["""a""", """b""", """c""", """d"""]} return Dataset.from_dict(lowerCamelCase ) def lowerCamelCase ( self : Union[str, Any] ) -> int: """simple docstring""" _UpperCAmelCase = self._create_example_records() _UpperCAmelCase = Dataset.from_list(lowerCamelCase ) self.assertListEqual(dset.column_names , ["""col_1""", """col_2"""] ) for i, r in enumerate(lowerCamelCase ): self.assertDictEqual(lowerCamelCase , example_records[i] ) def lowerCamelCase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = self._create_example_records() _UpperCAmelCase = Dataset.from_list(lowerCamelCase ) _UpperCAmelCase = Dataset.from_dict({k: [r[k] for r in example_records] for k in example_records[0]} ) self.assertEqual(dset.info , dset_from_dict.info ) def lowerCamelCase ( self : str ) -> Any: # checks what happens with missing columns """simple docstring""" _UpperCAmelCase = [{"""col_1""": 1}, {"""col_2""": """x"""}] _UpperCAmelCase = Dataset.from_list(lowerCamelCase ) self.assertDictEqual(dset[0] , {"""col_1""": 1} ) self.assertDictEqual(dset[1] , {"""col_1""": None} ) # NB: first record is used for columns def lowerCamelCase ( self : List[str] ) -> Optional[int]: # checks if the type can be inferred from the second record """simple docstring""" _UpperCAmelCase = [{"""col_1""": []}, {"""col_1""": [1, 2]}] _UpperCAmelCase = Dataset.from_list(lowerCamelCase ) self.assertEqual(dset.info.features["""col_1"""] , Sequence(Value("""int64""" ) ) ) def lowerCamelCase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" _UpperCAmelCase = Dataset.from_list([] ) self.assertEqual(len(lowerCamelCase ) , 0 ) self.assertListEqual(dset.column_names , [] )
108
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
0
'''simple docstring''' import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a = logging.get_logger(__name__) a = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } a = { "vocab_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json" }, "merges_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt" }, "tokenizer_config_file": { "facebook/blenderbot_small-90M": ( "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json" ) }, } a = {"facebook/blenderbot_small-90M": 512} def __magic_name__ ( __UpperCAmelCase ) -> Tuple: '''simple docstring''' __SCREAMING_SNAKE_CASE = set() __SCREAMING_SNAKE_CASE = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __SCREAMING_SNAKE_CASE = char __SCREAMING_SNAKE_CASE = set(__UpperCAmelCase ) return pairs class __a ( _snake_case ): __UpperCamelCase : Union[str, Any] = VOCAB_FILES_NAMES __UpperCamelCase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : List[str] = ['input_ids', 'attention_mask'] def __init__( self : int ,lowerCamelCase : str ,lowerCamelCase : int ,lowerCamelCase : Optional[int]="__start__" ,lowerCamelCase : Tuple="__end__" ,lowerCamelCase : int="__unk__" ,lowerCamelCase : Any="__null__" ,**lowerCamelCase : List[str] ,): '''simple docstring''' super().__init__(unk_token=lowerCamelCase ,bos_token=lowerCamelCase ,eos_token=lowerCamelCase ,pad_token=lowerCamelCase ,**lowerCamelCase ) with open(lowerCamelCase ,encoding="""utf-8""" ) as vocab_handle: __SCREAMING_SNAKE_CASE = json.load(lowerCamelCase ) __SCREAMING_SNAKE_CASE = {v: k for k, v in self.encoder.items()} with open(lowerCamelCase ,encoding="""utf-8""" ) as merges_handle: __SCREAMING_SNAKE_CASE = merges_handle.read().split("""\n""" )[1:-1] __SCREAMING_SNAKE_CASE = [tuple(merge.split() ) for merge in merges] __SCREAMING_SNAKE_CASE = dict(zip(lowerCamelCase ,range(len(lowerCamelCase ) ) ) ) __SCREAMING_SNAKE_CASE = {} @property def UpperCAmelCase__ ( self : Dict ): '''simple docstring''' return len(self.encoder ) def UpperCAmelCase__ ( self : str ): '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase__ ( self : List[Any] ,lowerCamelCase : str ): '''simple docstring''' if token in self.cache: return self.cache[token] __SCREAMING_SNAKE_CASE = re.sub("""([.,!?()])""" ,r""" \1""" ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = re.sub("""(')""" ,r""" \1 """ ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = re.sub(r"""\s{2,}""" ,""" """ ,lowerCamelCase ) if "\n" in token: __SCREAMING_SNAKE_CASE = token.replace("""\n""" ,""" __newln__""" ) __SCREAMING_SNAKE_CASE = token.split(""" """ ) __SCREAMING_SNAKE_CASE = [] for token in tokens: if not len(lowerCamelCase ): continue __SCREAMING_SNAKE_CASE = token.lower() __SCREAMING_SNAKE_CASE = tuple(lowerCamelCase ) __SCREAMING_SNAKE_CASE = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) __SCREAMING_SNAKE_CASE = get_pairs(lowerCamelCase ) if not pairs: words.append(lowerCamelCase ) continue while True: __SCREAMING_SNAKE_CASE = min(lowerCamelCase ,key=lambda lowerCamelCase : self.bpe_ranks.get(lowerCamelCase ,float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = bigram __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = 0 while i < len(lowerCamelCase ): try: __SCREAMING_SNAKE_CASE = word.index(lowerCamelCase ,lowerCamelCase ) new_word.extend(word[i:j] ) __SCREAMING_SNAKE_CASE = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(lowerCamelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __SCREAMING_SNAKE_CASE = tuple(lowerCamelCase ) __SCREAMING_SNAKE_CASE = new_word if len(lowerCamelCase ) == 1: break else: __SCREAMING_SNAKE_CASE = get_pairs(lowerCamelCase ) __SCREAMING_SNAKE_CASE = """@@ """.join(lowerCamelCase ) __SCREAMING_SNAKE_CASE = word[:-4] __SCREAMING_SNAKE_CASE = word words.append(lowerCamelCase ) return " ".join(lowerCamelCase ) def UpperCAmelCase__ ( self : int ,lowerCamelCase : str ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = re.findall(r"""\S+\n?""" ,lowerCamelCase ) for token in words: split_tokens.extend(list(self.bpe(lowerCamelCase ).split(""" """ ) ) ) return split_tokens def UpperCAmelCase__ ( self : Optional[Any] ,lowerCamelCase : str ): '''simple docstring''' __SCREAMING_SNAKE_CASE = token.lower() return self.encoder.get(lowerCamelCase ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase__ ( self : Tuple ,lowerCamelCase : int ): '''simple docstring''' return self.decoder.get(lowerCamelCase ,self.unk_token ) def UpperCAmelCase__ ( self : Union[str, Any] ,lowerCamelCase : List[str] ): '''simple docstring''' __SCREAMING_SNAKE_CASE = """ """.join(lowerCamelCase ).replace("""@@ """ ,"""""" ).strip() return out_string def UpperCAmelCase__ ( self : Optional[Any] ,lowerCamelCase : str ,lowerCamelCase : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(lowerCamelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __SCREAMING_SNAKE_CASE = os.path.join( lowerCamelCase ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __SCREAMING_SNAKE_CASE = os.path.join( lowerCamelCase ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(lowerCamelCase ,"""w""" ,encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCamelCase ,ensure_ascii=lowerCamelCase ) + """\n""" ) __SCREAMING_SNAKE_CASE = 0 with open(lowerCamelCase ,"""w""" ,encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCamelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" """ Please check that the tokenizer is not corrupted!""" ) __SCREAMING_SNAKE_CASE = token_index writer.write(""" """.join(lowerCamelCase ) + """\n""" ) index += 1 return vocab_file, merge_file
109
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
"""simple docstring""" import argparse import torch from transformers import YosoConfig, YosoForMaskedLM def lowerCamelCase ( _snake_case ): if "model" in orig_key: UpperCAmelCase__ : Union[str, Any] = orig_key.replace('model.' ,'' ) if "norm1" in orig_key: UpperCAmelCase__ : str = orig_key.replace('norm1' ,'attention.output.LayerNorm' ) if "norm2" in orig_key: UpperCAmelCase__ : Optional[int] = orig_key.replace('norm2' ,'output.LayerNorm' ) if "norm" in orig_key: UpperCAmelCase__ : Optional[int] = orig_key.replace('norm' ,'LayerNorm' ) if "transformer" in orig_key: UpperCAmelCase__ : Tuple = orig_key.split('.' )[0].split('_' )[-1] UpperCAmelCase__ : Optional[int] = orig_key.replace(F'''transformer_{layer_num}''' ,F'''encoder.layer.{layer_num}''' ) if "mha.attn" in orig_key: UpperCAmelCase__ : List[Any] = orig_key.replace('mha.attn' ,'attention.self' ) if "mha" in orig_key: UpperCAmelCase__ : int = orig_key.replace('mha' ,'attention' ) if "W_q" in orig_key: UpperCAmelCase__ : Dict = orig_key.replace('W_q' ,'self.query' ) if "W_k" in orig_key: UpperCAmelCase__ : Dict = orig_key.replace('W_k' ,'self.key' ) if "W_v" in orig_key: UpperCAmelCase__ : int = orig_key.replace('W_v' ,'self.value' ) if "ff1" in orig_key: UpperCAmelCase__ : int = orig_key.replace('ff1' ,'intermediate.dense' ) if "ff2" in orig_key: UpperCAmelCase__ : str = orig_key.replace('ff2' ,'output.dense' ) if "ff" in orig_key: UpperCAmelCase__ : Tuple = orig_key.replace('ff' ,'output.dense' ) if "mlm_class" in orig_key: UpperCAmelCase__ : Dict = orig_key.replace('mlm.mlm_class' ,'cls.predictions.decoder' ) if "mlm" in orig_key: UpperCAmelCase__ : Union[str, Any] = orig_key.replace('mlm' ,'cls.predictions.transform' ) if "cls" not in orig_key: UpperCAmelCase__ : Dict = 'yoso.' + orig_key return orig_key def lowerCamelCase ( _snake_case ,_snake_case ): for key in orig_state_dict.copy().keys(): UpperCAmelCase__ : Optional[int] = orig_state_dict.pop(_snake_case ) if ("pooler" in key) or ("sen_class" in key): continue else: UpperCAmelCase__ : Union[str, Any] = val UpperCAmelCase__ : Dict = orig_state_dict['cls.predictions.decoder.bias'] UpperCAmelCase__ : int = torch.arange(_snake_case ).expand((1, -1) ) + 2 return orig_state_dict def lowerCamelCase ( _snake_case ,_snake_case ,_snake_case ): UpperCAmelCase__ : Tuple = torch.load(_snake_case ,map_location='cpu' )['model_state_dict'] UpperCAmelCase__ : Union[str, Any] = YosoConfig.from_json_file(_snake_case ) UpperCAmelCase__ : int = YosoForMaskedLM(_snake_case ) UpperCAmelCase__ : Union[str, Any] = convert_checkpoint_helper(config.max_position_embeddings ,_snake_case ) print(model.load_state_dict(_snake_case ) ) model.eval() model.save_pretrained(_snake_case ) print(F'''Checkpoint successfuly converted. Model saved at {pytorch_dump_path}''' ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--pytorch_model_path', default=None, type=str, required=True, help='Path to YOSO pytorch checkpoint.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The json file for YOSO model config.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) UpperCamelCase__ = parser.parse_args() convert_yoso_checkpoint(args.pytorch_model_path, args.config_file, args.pytorch_dump_path)
110
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
0
'''simple docstring''' from ....utils import logging lowerCamelCase :List[str] = logging.get_logger(__name__) class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def __init__(self , lowercase , lowercase=None , lowercase=2048 ): A_ : Dict = config.__dict__ A_ : Tuple = modal_hidden_size if num_labels: A_ : Union[str, Any] = num_labels
667
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : Dict =KandinskyInpaintPipeline _UpperCAmelCase : str =["prompt", "image_embeds", "negative_image_embeds", "image", "mask_image"] _UpperCAmelCase : Any =[ "prompt", "negative_prompt", "image_embeds", "negative_image_embeds", "image", "mask_image", ] _UpperCAmelCase : Optional[Any] =[ "generator", "height", "width", "latents", "guidance_scale", "negative_prompt", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] _UpperCAmelCase : Dict =False @property def _UpperCAmelCase ( self : str ): return 32 @property def _UpperCAmelCase ( self : Dict ): return 32 @property def _UpperCAmelCase ( self : Any ): return self.time_input_dim @property def _UpperCAmelCase ( self : Optional[Any] ): return self.time_input_dim * 4 @property def _UpperCAmelCase ( self : Optional[Any] ): return 1_00 @property def _UpperCAmelCase ( self : Tuple ): A_ = XLMRobertaTokenizerFast.from_pretrained("YiYiXu/tiny-random-mclip-base" ) return tokenizer @property def _UpperCAmelCase ( self : Union[str, Any] ): torch.manual_seed(0 ) A_ = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=10_05 , ) A_ = MultilingualCLIP(snake_case_ ) A_ = text_encoder.eval() return text_encoder @property def _UpperCAmelCase ( self : Dict ): torch.manual_seed(0 ) A_ = { "in_channels": 9, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "text_image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "text_image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } A_ = UNetaDConditionModel(**snake_case_ ) return model @property def _UpperCAmelCase ( self : List[Any] ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def _UpperCAmelCase ( self : Union[str, Any] ): torch.manual_seed(0 ) A_ = VQModel(**self.dummy_movq_kwargs ) return model def _UpperCAmelCase ( self : Optional[Any] ): A_ = self.dummy_text_encoder A_ = self.dummy_tokenizer A_ = self.dummy_unet A_ = self.dummy_movq A_ = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule="linear" , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , steps_offset=1 , prediction_type="epsilon" , thresholding=snake_case_ , ) A_ = { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "movq": movq, } return components def _UpperCAmelCase ( self : Union[str, Any] , lowerCAmelCase : Optional[int] , lowerCAmelCase : str=0 ): A_ = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A_ = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(snake_case_ ) # create init_image A_ = floats_tensor((1, 3, 64, 64) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A_ = image.cpu().permute(0 , 2 , 3 , 1 )[0] A_ = Image.fromarray(np.uinta(snake_case_ ) ).convert("RGB" ).resize((2_56, 2_56) ) # create mask A_ = np.ones((64, 64) , dtype=np.floataa ) A_ = 0 if str(snake_case_ ).startswith("mps" ): A_ = torch.manual_seed(snake_case_ ) else: A_ = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) A_ = { "prompt": "horse", "image": init_image, "mask_image": mask, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 2, "guidance_scale": 4.0, "output_type": "np", } return inputs def _UpperCAmelCase ( self : Optional[Any] ): A_ = "cpu" A_ = self.get_dummy_components() A_ = self.pipeline_class(**snake_case_ ) A_ = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) A_ = pipe(**self.get_dummy_inputs(snake_case_ ) ) A_ = output.images A_ = pipe( **self.get_dummy_inputs(snake_case_ ) , return_dict=snake_case_ , )[0] A_ = image[0, -3:, -3:, -1] A_ = image_from_tuple[0, -3:, -3:, -1] print(F"image.shape {image.shape}" ) assert image.shape == (1, 64, 64, 3) A_ = np.array( [0.8_3_2_6_9_1_9, 0.7_3_7_9_0_4_6_7, 0.2_0_9_1_8_5_8_1, 0.9_3_0_9_6_1_2, 0.5_5_1_1_7_9_1, 0.4_3_7_1_3_3_2_8, 0.5_5_1_3_3_2_1, 0.4_9_9_2_2_9_3_4, 0.5_9_4_9_7_7_8_6] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" def _UpperCAmelCase ( self : str ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _UpperCAmelCase ( self : List[Any] ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCAmelCase ( self : Optional[Any] ): A_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy" ) A_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) A_ = np.ones((7_68, 7_68) , dtype=np.floataa ) A_ = 0 A_ = "a hat" A_ = KandinskyPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-1-prior" , torch_dtype=torch.floataa ) pipe_prior.to(snake_case_ ) A_ = KandinskyInpaintPipeline.from_pretrained( "kandinsky-community/kandinsky-2-1-inpaint" , torch_dtype=torch.floataa ) A_ = pipeline.to(snake_case_ ) pipeline.set_progress_bar_config(disable=snake_case_ ) A_ = torch.Generator(device="cpu" ).manual_seed(0 ) A_ , A_ = pipe_prior( snake_case_ , generator=snake_case_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() A_ = pipeline( snake_case_ , image=snake_case_ , mask_image=snake_case_ , image_embeds=snake_case_ , negative_image_embeds=snake_case_ , generator=snake_case_ , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type="np" , ) A_ = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(snake_case_ , snake_case_ )
452
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
0
"""simple docstring""" import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> Any: assert isinstance(lowercase_, lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory", [False, True] ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> Tuple: a_ : Optional[Any] = tmp_path / "cache" a_ : Union[str, Any] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): a_ : int = ParquetDatasetReader(lowercase_, cache_dir=lowercase_, keep_in_memory=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) @pytest.mark.parametrize( "features", [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ], ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> int: a_ : int = tmp_path / "cache" a_ : Tuple = {"col_1": "string", "col_2": "int64", "col_3": "float64"} a_ : str = features.copy() if features else default_expected_features a_ : List[Any] = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) a_ : List[str] = ParquetDatasetReader(lowercase_, features=lowercase_, cache_dir=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) @pytest.mark.parametrize("split", [None, NamedSplit("train" ), "train", "test"] ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> Optional[int]: a_ : Any = tmp_path / "cache" a_ : Tuple = {"col_1": "string", "col_2": "int64", "col_3": "float64"} a_ : str = ParquetDatasetReader(lowercase_, cache_dir=lowercase_, split=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type", [str, list] ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> Any: if issubclass(lowercase_, lowercase_ ): a_ : List[Any] = parquet_path elif issubclass(lowercase_, lowercase_ ): a_ : str = [parquet_path] a_ : Dict = tmp_path / "cache" a_ : List[str] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} a_ : List[Any] = ParquetDatasetReader(lowercase_, cache_dir=lowercase_ ).read() _check_parquet_dataset(lowercase_, lowercase_ ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__=("train",) ) -> Optional[Any]: assert isinstance(lowercase_, lowercase_ ) for split in splits: a_ : str = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory", [False, True] ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> int: a_ : Dict = tmp_path / "cache" a_ : Any = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): a_ : List[str] = ParquetDatasetReader( {"train": parquet_path}, cache_dir=lowercase_, keep_in_memory=lowercase_ ).read() _check_parquet_datasetdict(lowercase_, lowercase_ ) @pytest.mark.parametrize( "features", [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ], ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> List[str]: a_ : str = tmp_path / "cache" a_ : Optional[int] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} a_ : str = features.copy() if features else default_expected_features a_ : Optional[int] = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) a_ : Optional[Any] = ParquetDatasetReader({"train": parquet_path}, features=lowercase_, cache_dir=lowercase_ ).read() _check_parquet_datasetdict(lowercase_, lowercase_ ) @pytest.mark.parametrize("split", [None, NamedSplit("train" ), "train", "test"] ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> Optional[int]: if split: a_ : Union[str, Any] = {split: parquet_path} else: a_ : Any = "train" a_ : Optional[int] = {"train": parquet_path, "test": parquet_path} a_ : Optional[int] = tmp_path / "cache" a_ : int = {"col_1": "string", "col_2": "int64", "col_3": "float64"} a_ : List[str] = ParquetDatasetReader(lowercase_, cache_dir=lowercase_ ).read() _check_parquet_datasetdict(lowercase_, lowercase_, splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> Optional[int]: a_ : Optional[int] = ParquetDatasetWriter(lowercase_, tmp_path / "foo.parquet" ) assert writer.write() > 0 a_ : int = pq.ParquetFile(tmp_path / "foo.parquet" ) a_ : List[Any] = pf.read() assert dataset.data.table == output_table def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> str: a_ : Dict = str(shared_datadir / "test_image_rgb.jpg" ) a_ : str = {"image": [image_path]} a_ : int = Features({"image": Image()} ) a_ : List[Any] = Dataset.from_dict(lowercase_, features=lowercase_ ) a_ : int = ParquetDatasetWriter(lowercase_, tmp_path / "foo.parquet" ) assert writer.write() > 0 a_ : int = Dataset.from_parquet(str(tmp_path / "foo.parquet" ) ) assert dataset.features == reloaded_dataset.features a_ : str = ParquetDatasetReader(str(tmp_path / "foo.parquet" ), streaming=lowercase_ ).read() assert dataset.features == reloaded_iterable_dataset.features @pytest.mark.parametrize( "feature, expected", [ (Features({"foo": Value("int32" )} ), None), (Features({"image": Image(), "foo": Value("int32" )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS), (Features({"nested": Sequence(Audio() )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS), ], ) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) -> List[Any]: assert get_writer_batch_size(lowercase_ ) == expected
237
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
0
"""simple docstring""" import pytest import datasets # Import fixture modules as plugins UpperCamelCase = ['''tests.fixtures.files''', '''tests.fixtures.hub''', '''tests.fixtures.fsspec'''] def _lowerCamelCase ( UpperCAmelCase_ : List[Any], UpperCAmelCase_ : Dict ) -> Tuple: """simple docstring""" for item in items: if any(marker in item.keywords for marker in ["integration", "unit"] ): continue item.add_marker(pytest.mark.unit ) def _lowerCamelCase ( UpperCAmelCase_ : Optional[Any] ) -> Optional[Any]: """simple docstring""" config.addinivalue_line("markers", "torchaudio_latest: mark test to run with torchaudio>=0.12" ) @pytest.fixture(autouse=lowercase_ ) def _lowerCamelCase ( UpperCAmelCase_ : List[str], UpperCAmelCase_ : Optional[int] ) -> str: """simple docstring""" A__ = tmp_path_factory.getbasetemp() / "cache" A__ = test_hf_cache_home / "datasets" A__ = test_hf_cache_home / "metrics" A__ = test_hf_cache_home / "modules" monkeypatch.setattr("datasets.config.HF_DATASETS_CACHE", str(lowercase_ ) ) monkeypatch.setattr("datasets.config.HF_METRICS_CACHE", str(lowercase_ ) ) monkeypatch.setattr("datasets.config.HF_MODULES_CACHE", str(lowercase_ ) ) A__ = test_hf_datasets_cache / "downloads" monkeypatch.setattr("datasets.config.DOWNLOADED_DATASETS_PATH", str(lowercase_ ) ) A__ = test_hf_datasets_cache / "downloads" / "extracted" monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH", str(lowercase_ ) ) @pytest.fixture(autouse=lowercase_, scope="session" ) def _lowerCamelCase ( ) -> Dict: """simple docstring""" datasets.disable_progress_bar() @pytest.fixture(autouse=lowercase_ ) def _lowerCamelCase ( UpperCAmelCase_ : List[Any] ) -> Any: """simple docstring""" monkeypatch.setattr("datasets.config.HF_UPDATE_DOWNLOAD_COUNTS", lowercase_ ) @pytest.fixture def _lowerCamelCase ( UpperCAmelCase_ : str ) -> Optional[Any]: """simple docstring""" monkeypatch.setattr("sqlalchemy.util.deprecations.SILENCE_UBER_WARNING", lowercase_ )
104
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
0
"""simple docstring""" import doctest from collections import deque import numpy as np class lowerCamelCase : '''simple docstring''' def __init__(self ): """simple docstring""" UpperCAmelCase__ : int = [2, 1, 2, -1] UpperCAmelCase__ : Dict = [1, 2, 3, 4] def _a (self ): """simple docstring""" UpperCAmelCase__ : str = len(self.first_signal ) UpperCAmelCase__ : Any = len(self.second_signal ) UpperCAmelCase__ : Tuple = max(snake_case_ , snake_case_ ) # create a zero matrix of max_length x max_length UpperCAmelCase__ : Any = [[0] * max_length for i in range(snake_case_ )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(snake_case_ ): UpperCAmelCase__ : Tuple = deque(self.second_signal ) rotated_signal.rotate(snake_case_ ) for j, item in enumerate(snake_case_ ): matrix[i][j] += item # multiply the matrix with the first signal UpperCAmelCase__ : Optional[Any] = np.matmul(np.transpose(snake_case_ ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(snake_case_ , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
182
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
0
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import InterpolationMode import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, ViTImageProcessor, ViTMAEConfig, ViTMAEForPreTraining, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version SCREAMING_SNAKE_CASE : str = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt") @dataclass class _lowerCamelCase: lowercase_ : Union[str, Any] = field( default="""cifar10""", metadata={"""help""": """Name of a dataset from the datasets package"""} ) lowercase_ : int = field( default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) lowercase_ : str = field( default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """The column name of the images in the files."""} ) lowercase_ : Tuple = field(default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """A folder containing the training data."""} ) lowercase_ : Dict = field(default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """A folder containing the validation data."""} ) lowercase_ : List[str] = field( default=0.15, metadata={"""help""": """Percent to split off of train for validation."""} ) lowercase_ : List[str] = field( default=__SCREAMING_SNAKE_CASE, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) lowercase_ : Any = field( default=__SCREAMING_SNAKE_CASE, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : List[str] = {} if self.train_dir is not None: _lowercase : Dict = self.train_dir if self.validation_dir is not None: _lowercase : Tuple = self.validation_dir _lowercase : List[Any] = data_files if data_files else None @dataclass class _lowerCamelCase: lowercase_ : Dict = field( default=__SCREAMING_SNAKE_CASE, metadata={ """help""": ( """The model checkpoint for weights initialization.Don\'t set if you want to train a model from scratch.""" ) }, ) lowercase_ : Any = field( default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """Pretrained config name or path if not the same as model_name_or_path"""} ) lowercase_ : str = field( default=__SCREAMING_SNAKE_CASE, metadata={ """help""": ( """Override some existing default config settings when a model is trained from scratch. Example: """ """n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index""" ) }, ) lowercase_ : List[str] = field( default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """Where do you want to store the pretrained models downloaded from s3"""} ) lowercase_ : Union[str, Any] = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) lowercase_ : Optional[int] = field(default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """Name or path of preprocessor config."""} ) lowercase_ : List[str] = field( default=__SCREAMING_SNAKE_CASE, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) lowercase_ : Union[str, Any] = field( default=0.75, metadata={"""help""": """The ratio of the number of masked tokens in the input sequence."""} ) lowercase_ : List[Any] = field( default=__SCREAMING_SNAKE_CASE, metadata={"""help""": """Whether or not to train with normalized pixel values as target."""} ) @dataclass class _lowerCamelCase( __SCREAMING_SNAKE_CASE ): lowercase_ : Tuple = field( default=1e-3, metadata={"""help""": """Base learning rate: absolute_lr = base_lr * total_batch_size / 256."""} ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : List[Any] = torch.stack([example['pixel_values'] for example in examples] ) return {"pixel_values": pixel_values} def UpperCamelCase_( ) -> Tuple: _lowercase : Union[str, Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, CustomTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowercase , _lowercase , _lowercase : List[Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowercase , _lowercase , _lowercase : List[str] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_mae' , lowercase_ , lowercase_ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _lowercase : int = training_args.get_process_log_level() logger.setLevel(lowercase_ ) transformers.utils.logging.set_verbosity(lowercase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. _lowercase : Tuple = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowercase : Union[str, Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Initialize our dataset. _lowercase : List[str] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. _lowercase : Optional[int] = None if 'validation' in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , lowercase_ ) and data_args.train_val_split > 0.0: _lowercase : int = ds['train'].train_test_split(data_args.train_val_split ) _lowercase : Tuple = split['train'] _lowercase : Dict = split['test'] # Load pretrained model and image processor # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : Any = { 'cache_dir': model_args.cache_dir, 'revision': model_args.model_revision, 'use_auth_token': True if model_args.use_auth_token else None, } if model_args.config_name: _lowercase : List[str] = ViTMAEConfig.from_pretrained(model_args.config_name , **lowercase_ ) elif model_args.model_name_or_path: _lowercase : List[Any] = ViTMAEConfig.from_pretrained(model_args.model_name_or_path , **lowercase_ ) else: _lowercase : Tuple = ViTMAEConfig() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.config_overrides is not None: logger.info(F'''Overriding config: {model_args.config_overrides}''' ) config.update_from_string(model_args.config_overrides ) logger.info(F'''New config: {config}''' ) # adapt config config.update( { 'mask_ratio': model_args.mask_ratio, 'norm_pix_loss': model_args.norm_pix_loss, } ) # create image processor if model_args.image_processor_name: _lowercase : str = ViTImageProcessor.from_pretrained(model_args.image_processor_name , **lowercase_ ) elif model_args.model_name_or_path: _lowercase : str = ViTImageProcessor.from_pretrained(model_args.model_name_or_path , **lowercase_ ) else: _lowercase : Union[str, Any] = ViTImageProcessor() # create model if model_args.model_name_or_path: _lowercase : Union[str, Any] = ViTMAEForPreTraining.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowercase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info('Training new model from scratch' ) _lowercase : Optional[int] = ViTMAEForPreTraining(lowercase_ ) if training_args.do_train: _lowercase : str = ds['train'].column_names else: _lowercase : Optional[int] = ds['validation'].column_names if data_args.image_column_name is not None: _lowercase : int = data_args.image_column_name elif "image" in column_names: _lowercase : List[Any] = 'image' elif "img" in column_names: _lowercase : List[Any] = 'img' else: _lowercase : int = column_names[0] # transformations as done in original MAE paper # source: https://github.com/facebookresearch/mae/blob/main/main_pretrain.py if "shortest_edge" in image_processor.size: _lowercase : str = image_processor.size['shortest_edge'] else: _lowercase : List[Any] = (image_processor.size['height'], image_processor.size['width']) _lowercase : Union[str, Any] = Compose( [ Lambda(lambda lowerCamelCase_ : img.convert('RGB' ) if img.mode != "RGB" else img ), RandomResizedCrop(lowercase_ , scale=(0.2, 1.0) , interpolation=InterpolationMode.BICUBIC ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) def preprocess_images(lowerCamelCase_ ): _lowercase : Optional[int] = [transforms(lowercase_ ) for image in examples[image_column_name]] return examples if training_args.do_train: if "train" not in ds: raise ValueError('--do_train requires a train dataset' ) if data_args.max_train_samples is not None: _lowercase : Union[str, Any] = ds['train'].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(lowercase_ ) if training_args.do_eval: if "validation" not in ds: raise ValueError('--do_eval requires a validation dataset' ) if data_args.max_eval_samples is not None: _lowercase : List[Any] = ( ds['validation'].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(lowercase_ ) # Compute absolute learning rate _lowercase : Union[str, Any] = ( training_args.train_batch_size * training_args.gradient_accumulation_steps * training_args.world_size ) if training_args.base_learning_rate is not None: _lowercase : Any = training_args.base_learning_rate * total_train_batch_size / 256 # Initialize our trainer _lowercase : List[Any] = Trainer( model=lowercase_ , args=lowercase_ , train_dataset=ds['train'] if training_args.do_train else None , eval_dataset=ds['validation'] if training_args.do_eval else None , tokenizer=lowercase_ , data_collator=lowercase_ , ) # Training if training_args.do_train: _lowercase : Any = None if training_args.resume_from_checkpoint is not None: _lowercase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowercase : Any = last_checkpoint _lowercase : Optional[int] = trainer.train(resume_from_checkpoint=lowercase_ ) trainer.save_model() trainer.log_metrics('train' , train_result.metrics ) trainer.save_metrics('train' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _lowercase : Any = trainer.evaluate() trainer.log_metrics('eval' , lowercase_ ) trainer.save_metrics('eval' , lowercase_ ) # Write model card and (optionally) push to hub _lowercase : Optional[int] = { 'tasks': 'masked-auto-encoding', 'dataset': data_args.dataset_name, 'tags': ['masked-auto-encoding'], } if training_args.push_to_hub: trainer.push_to_hub(**lowercase_ ) else: trainer.create_model_card(**lowercase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: main() if __name__ == "__main__": main()
89
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
0
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging _lowerCamelCase = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def __UpperCAmelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_=None ): _lowerCamelCase : Optional[int] = XLNetConfig.from_json_file(lowercase_ ) _lowerCamelCase : Any = finetuning_task.lower() if finetuning_task is not None else '''''' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(F"""Building PyTorch XLNetForSequenceClassification model from configuration: {config}""" ) _lowerCamelCase : Dict = finetuning_task _lowerCamelCase : List[str] = GLUE_TASKS_NUM_LABELS[finetuning_task] _lowerCamelCase : str = XLNetForSequenceClassification(lowercase_ ) elif "squad" in finetuning_task: _lowerCamelCase : List[str] = finetuning_task _lowerCamelCase : Optional[int] = XLNetForQuestionAnswering(lowercase_ ) else: _lowerCamelCase : Tuple = XLNetLMHeadModel(lowercase_ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(lowercase_ , lowercase_ , lowercase_ ) # Save pytorch-model _lowerCamelCase : Union[str, Any] = os.path.join(lowercase_ , lowercase_ ) _lowerCamelCase : Tuple = os.path.join(lowercase_ , lowercase_ ) print(F"""Save PyTorch model to {os.path.abspath(lowercase_ )}""" ) torch.save(model.state_dict() , lowercase_ ) print(F"""Save configuration file to {os.path.abspath(lowercase_ )}""" ) with open(lowercase_ , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--xlnet_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained XLNet model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--finetuning_task', default=None, type=str, help='Name of a task on which the XLNet TensorFlow model was fine-tuned', ) _lowerCamelCase = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
114
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
0
def UpperCamelCase ( __lowercase : int ): '''simple docstring''' A_ : Tuple = (1 + 24 * n) ** 0.5 return ((1 + root) / 6) % 1 == 0 def UpperCamelCase ( __lowercase : int = 50_00 ): '''simple docstring''' A_ : str = [(i * (3 * i - 1)) // 2 for i in range(1 ,lowercase_ )] for i, pentagonal_i in enumerate(lowercase_ ): for j in range(lowercase_ ,len(lowercase_ ) ): A_ : Optional[int] = pentagonal_nums[j] A_ : Optional[int] = pentagonal_i + pentagonal_j A_ : Any = pentagonal_j - pentagonal_i if is_pentagonal(lowercase_ ) and is_pentagonal(lowercase_ ): return b return -1 if __name__ == "__main__": print(F"""{solution() = }""")
558
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
0
'''simple docstring''' def _UpperCAmelCase ( _lowerCamelCase : int = 50 ) -> int: _lowerCAmelCase : List[Any] = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F'{solution() = }')
384
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
0
"""simple docstring""" from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def UpperCAmelCase ( snake_case : List[Any] , snake_case : Tuple , snake_case : int , snake_case : Optional[Any] ): for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), F'Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})' else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), F'Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})' def UpperCAmelCase ( snake_case : int , snake_case : Any , snake_case : Union[str, Any] , snake_case : Optional[int] , snake_case : Union[str, Any]=True ): model.train() _lowerCAmelCase:Union[str, Any] = model(lowercase_ ) _lowerCAmelCase:str = F.mse_loss(lowercase_ , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowercase_ ) def UpperCAmelCase ( snake_case : str , snake_case : Dict=False ): set_seed(42 ) _lowerCAmelCase:int = RegressionModel() _lowerCAmelCase:List[Any] = deepcopy(lowercase_ ) _lowerCAmelCase:Union[str, Any] = RegressionDataset(length=80 ) _lowerCAmelCase:Optional[Any] = DataLoader(lowercase_ , batch_size=16 ) model.to(accelerator.device ) if sched: _lowerCAmelCase:Any = AdamW(params=model.parameters() , lr=1e-3 ) _lowerCAmelCase:Any = AdamW(params=ddp_model.parameters() , lr=1e-3 ) _lowerCAmelCase:int = LambdaLR(lowercase_ , lr_lambda=lambda snake_case : epoch**0.65 ) _lowerCAmelCase:Optional[Any] = LambdaLR(lowercase_ , lr_lambda=lambda snake_case : epoch**0.65 ) # Make a copy of `model` if sched: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase:Optional[Any] = accelerator.prepare(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: _lowerCAmelCase , _lowerCAmelCase:Dict = accelerator.prepare(lowercase_ , lowercase_ ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def UpperCAmelCase ( snake_case : int ): _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase:int = get_training_setup(lowercase_ ) # Use a single batch _lowerCAmelCase , _lowerCAmelCase:Optional[Any] = next(iter(lowercase_ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model _lowerCAmelCase , _lowerCAmelCase:str = accelerator.gather((ddp_input, ddp_target) ) _lowerCAmelCase , _lowerCAmelCase:str = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowercase_ ): step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: # Sync grads step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), F'Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) _lowerCAmelCase:int = ddp_input[torch.randperm(len(lowercase_ ) )] def UpperCAmelCase ( snake_case : Any ): _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase:Tuple = get_training_setup(lowercase_ ) # Use a single batch _lowerCAmelCase , _lowerCAmelCase:Optional[int] = next(iter(lowercase_ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model _lowerCAmelCase , _lowerCAmelCase:Optional[int] = accelerator.gather((ddp_input, ddp_target) ) _lowerCAmelCase , _lowerCAmelCase:Union[str, Any] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowercase_ ): step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: # Sync grads step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})' else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) _lowerCAmelCase:Tuple = ddp_input[torch.randperm(len(lowercase_ ) )] def UpperCAmelCase ( snake_case : Dict=False , snake_case : Tuple=False ): _lowerCAmelCase:int = Accelerator( split_batches=lowercase_ , dispatch_batches=lowercase_ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase:Union[str, Any] = get_training_setup(lowercase_ ) for iteration, batch in enumerate(lowercase_ ): _lowerCAmelCase , _lowerCAmelCase:Optional[int] = batch.values() # Gather the distributed inputs and targs for the base model _lowerCAmelCase , _lowerCAmelCase:Optional[int] = accelerator.gather((ddp_input, ddp_target) ) _lowerCAmelCase , _lowerCAmelCase:List[Any] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowercase_ ): step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowercase_ ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F'Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})' else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F'Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})' # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) _lowerCAmelCase:Union[str, Any] = ddp_input[torch.randperm(len(lowercase_ ) )] GradientState._reset_state() def UpperCAmelCase ( snake_case : List[str]=False , snake_case : str=False ): _lowerCAmelCase:str = Accelerator( split_batches=lowercase_ , dispatch_batches=lowercase_ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase:Dict = get_training_setup(lowercase_ , lowercase_ ) for iteration, batch in enumerate(lowercase_ ): _lowerCAmelCase , _lowerCAmelCase:Union[str, Any] = batch.values() # Gather the distributed inputs and targs for the base model _lowerCAmelCase , _lowerCAmelCase:str = accelerator.gather((ddp_input, ddp_target) ) _lowerCAmelCase , _lowerCAmelCase:Dict = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowercase_ )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowercase_ ): step_model(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F'Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n' _lowerCAmelCase:int = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowercase_ )) if accelerator.num_processes > 1: check_model_parameters(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def UpperCAmelCase ( ): _lowerCAmelCase:List[str] = Accelerator() _lowerCAmelCase:Optional[int] = RegressionDataset(length=80 ) _lowerCAmelCase:str = DataLoader(lowercase_ , batch_size=16 ) _lowerCAmelCase:List[Any] = RegressionDataset(length=96 ) _lowerCAmelCase:List[Any] = DataLoader(lowercase_ , batch_size=16 ) _lowerCAmelCase , _lowerCAmelCase:Dict = accelerator.prepare(lowercase_ , lowercase_ ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowercase_ ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowercase_ ) if iteration < len(lowercase_ ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowercase_ ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowercase_ ) if batch_num < len(lowercase_ ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def UpperCAmelCase ( ): _lowerCAmelCase:Any = Accelerator() _lowerCAmelCase:int = accelerator.state if state.local_process_index == 0: print('''**Test `accumulate` gradient accumulation with dataloader break**''' ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print('''**Test NOOP `no_sync` context manager**''' ) test_noop_sync(lowercase_ ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print('''**Test Distributed `no_sync` context manager**''' ) test_distributed_sync(lowercase_ ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( '''**Test `accumulate` gradient accumulation, ''' , F'`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**' , ) test_gradient_accumulation(lowercase_ , lowercase_ ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version('''<''' , '''2.0''' ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( '''**Test `accumulate` gradient accumulation with optimizer and scheduler, ''' , '''`split_batches=False`, `dispatch_batches=False`**''' , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( '''**Test `accumulate` gradient accumulation with optimizer and scheduler, ''' , F'`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**' , ) test_gradient_accumulation_with_opt_and_scheduler(lowercase_ , lowercase_ ) def UpperCAmelCase ( snake_case : Dict ): main() if __name__ == "__main__": main()
227
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Optional[Any] = logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case__ = """encoder-decoder""" snake_case__ = True def __init__( self : List[str] , **_SCREAMING_SNAKE_CASE : int ): """simple docstring""" super().__init__(**snake_case_ ) assert ( "encoder" in kwargs and "decoder" in kwargs ), "Config has to be initialized with encoder and decoder config" UpperCamelCase = kwargs.pop('encoder' ) UpperCamelCase = encoder_config.pop('model_type' ) UpperCamelCase = kwargs.pop('decoder' ) UpperCamelCase = decoder_config.pop('model_type' ) from ..auto.configuration_auto import AutoConfig UpperCamelCase = AutoConfig.for_model(snake_case_ , **snake_case_ ) UpperCamelCase = AutoConfig.for_model(snake_case_ , **snake_case_ ) UpperCamelCase = True @classmethod def _SCREAMING_SNAKE_CASE ( cls : int , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : str , **_SCREAMING_SNAKE_CASE : Tuple ): """simple docstring""" logger.info('Set `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config' ) UpperCamelCase = True UpperCamelCase = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **snake_case_ ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = copy.deepcopy(self.__dict__ ) UpperCamelCase = self.encoder.to_dict() UpperCamelCase = self.decoder.to_dict() UpperCamelCase = self.__class__.model_type return output
280
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
0
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : List[Any] = ['image_processor'] __SCREAMING_SNAKE_CASE : List[str] = 'SamImageProcessor' def __init__(self , lowercase ): super().__init__(snake_case_ ) A_ : Dict = self.image_processor A_ : Any = -10 A_ : Optional[int] = self.image_processor.size["""longest_edge"""] def __call__(self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase = None , **lowercase , ): A_ : List[Any] = self.image_processor( snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # pop arguments that are not used in the foward but used nevertheless A_ : Tuple = encoding_image_processor["""original_sizes"""] if hasattr(snake_case_ , """numpy""" ): # Checks if Torch or TF tensor A_ : Any = original_sizes.numpy() A_, A_, A_ : str = self._check_and_preprocess_points( input_points=snake_case_ , input_labels=snake_case_ , input_boxes=snake_case_ , ) A_ : int = self._normalize_and_convert( snake_case_ , snake_case_ , input_points=snake_case_ , input_labels=snake_case_ , input_boxes=snake_case_ , return_tensors=snake_case_ , ) return encoding_image_processor def _a (self , lowercase , lowercase , lowercase=None , lowercase=None , lowercase=None , lowercase="pt" , ): if input_points is not None: if len(snake_case_ ) != len(snake_case_ ): A_ : Optional[int] = [ self._normalize_coordinates(self.target_size , snake_case_ , original_sizes[0] ) for point in input_points ] else: A_ : str = [ self._normalize_coordinates(self.target_size , snake_case_ , snake_case_ ) for point, original_size in zip(snake_case_ , snake_case_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: A_, A_ : Optional[Any] = self._pad_points_and_labels(snake_case_ , snake_case_ ) A_ : Optional[int] = np.array(snake_case_ ) if input_labels is not None: A_ : Union[str, Any] = np.array(snake_case_ ) if input_boxes is not None: if len(snake_case_ ) != len(snake_case_ ): A_ : List[str] = [ self._normalize_coordinates(self.target_size , snake_case_ , original_sizes[0] , is_bounding_box=snake_case_ ) for box in input_boxes ] else: A_ : Union[str, Any] = [ self._normalize_coordinates(self.target_size , snake_case_ , snake_case_ , is_bounding_box=snake_case_ ) for box, original_size in zip(snake_case_ , snake_case_ ) ] A_ : List[Any] = np.array(snake_case_ ) if input_boxes is not None: if return_tensors == "pt": A_ : Optional[int] = torch.from_numpy(snake_case_ ) # boxes batch size of 1 by default A_ : List[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": A_ : int = tf.convert_to_tensor(snake_case_ ) # boxes batch size of 1 by default A_ : Any = tf.expand_dims(snake_case_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"""input_boxes""": input_boxes} ) if input_points is not None: if return_tensors == "pt": A_ : List[Any] = torch.from_numpy(snake_case_ ) # point batch size of 1 by default A_ : Optional[int] = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": A_ : str = tf.convert_to_tensor(snake_case_ ) # point batch size of 1 by default A_ : Union[str, Any] = tf.expand_dims(snake_case_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"""input_points""": input_points} ) if input_labels is not None: if return_tensors == "pt": A_ : Union[str, Any] = torch.from_numpy(snake_case_ ) # point batch size of 1 by default A_ : Tuple = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": A_ : Optional[Any] = tf.convert_to_tensor(snake_case_ ) # point batch size of 1 by default A_ : int = tf.expand_dims(snake_case_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"""input_labels""": input_labels} ) return encoding_image_processor def _a (self , lowercase , lowercase ): A_ : Tuple = max([point.shape[0] for point in input_points] ) A_ : Optional[Any] = [] for i, point in enumerate(snake_case_ ): if point.shape[0] != expected_nb_points: A_ : Optional[int] = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) A_ : Optional[int] = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(snake_case_ ) A_ : int = processed_input_points return input_points, input_labels def _a (self , lowercase , lowercase , lowercase , lowercase=False ): A_, A_ : Optional[int] = original_size A_, A_ : Optional[int] = self.image_processor._get_preprocess_shape(snake_case_ , longest_edge=snake_case_ ) A_ : int = deepcopy(snake_case_ ).astype(snake_case_ ) if is_bounding_box: A_ : Any = coords.reshape(-1 , 2 , 2 ) A_ : str = coords[..., 0] * (new_w / old_w) A_ : str = coords[..., 1] * (new_h / old_h) if is_bounding_box: A_ : Optional[Any] = coords.reshape(-1 , 4 ) return coords def _a (self , lowercase=None , lowercase=None , lowercase=None , ): if input_points is not None: if hasattr(snake_case_ , """numpy""" ): # Checks for TF or Torch tensor A_ : Union[str, Any] = input_points.numpy().tolist() if not isinstance(snake_case_ , snake_case_ ) or not isinstance(input_points[0] , snake_case_ ): raise ValueError("""Input points must be a list of list of floating points.""" ) A_ : Any = [np.array(snake_case_ ) for input_point in input_points] else: A_ : int = None if input_labels is not None: if hasattr(snake_case_ , """numpy""" ): A_ : Dict = input_labels.numpy().tolist() if not isinstance(snake_case_ , snake_case_ ) or not isinstance(input_labels[0] , snake_case_ ): raise ValueError("""Input labels must be a list of list integers.""" ) A_ : Dict = [np.array(snake_case_ ) for label in input_labels] else: A_ : List[str] = None if input_boxes is not None: if hasattr(snake_case_ , """numpy""" ): A_ : Union[str, Any] = input_boxes.numpy().tolist() if ( not isinstance(snake_case_ , snake_case_ ) or not isinstance(input_boxes[0] , snake_case_ ) or not isinstance(input_boxes[0][0] , snake_case_ ) ): raise ValueError("""Input boxes must be a list of list of list of floating points.""" ) A_ : List[str] = [np.array(snake_case_ ).astype(np.floataa ) for box in input_boxes] else: A_ : str = None return input_points, input_labels, input_boxes @property def _a (self ): A_ : Optional[Any] = self.image_processor.model_input_names return list(dict.fromkeys(snake_case_ ) ) def _a (self , *lowercase , **lowercase ): return self.image_processor.post_process_masks(*snake_case_ , **snake_case_ )
667
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor __SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): """simple docstring""" def __init__( self : Optional[int] , *lowerCAmelCase : Tuple , **lowerCAmelCase : Union[str, Any] ): warnings.warn( "The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use YolosImageProcessor instead." , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
452
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
0
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = {'''vocab_file''': '''spiece.model'''} SCREAMING_SNAKE_CASE_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } SCREAMING_SNAKE_CASE_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = 3 SCREAMING_SNAKE_CASE_ = 4 class snake_case_ ( __SCREAMING_SNAKE_CASE ): __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = "left" def __init__( self , a_ , a_=False , a_=True , a_=False , a_="<s>" , a_="</s>" , a_="<unk>" , a_="<sep>" , a_="<pad>" , a_="<cls>" , a_="<mask>" , a_=["<eop>", "<eod>"] , a_ = None , **a_ , ): # Mask token behave like a normal word, i.e. include the space before it a_ : int = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token a_ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case_ , remove_space=snake_case_ , keep_accents=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , pad_token=snake_case_ , cls_token=snake_case_ , mask_token=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) a_ : Optional[int] = 3 a_ : str = do_lower_case a_ : int = remove_space a_ : Dict = keep_accents a_ : Tuple = vocab_file a_ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) @property def snake_case_ ( self ): return len(self.sp_model ) def snake_case_ ( self ): a_ : Dict = {self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): a_ : List[Any] = self.__dict__.copy() a_ : Optional[int] = None return state def __setstate__( self , a_ ): a_ : List[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): a_ : str = {} a_ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def snake_case_ ( self , a_ ): if self.remove_space: a_ : Optional[int] = " ".join(inputs.strip().split() ) else: a_ : List[str] = inputs a_ : Optional[Any] = outputs.replace("``" , "\"" ).replace("\'\'" , "\"" ) if not self.keep_accents: a_ : Dict = unicodedata.normalize("NFKD" , snake_case_ ) a_ : int = "".join([c for c in outputs if not unicodedata.combining(snake_case_ )] ) if self.do_lower_case: a_ : Union[str, Any] = outputs.lower() return outputs def snake_case_ ( self , a_ ): a_ : int = self.preprocess_text(snake_case_ ) a_ : Optional[int] = self.sp_model.encode(snake_case_ , out_type=snake_case_ ) a_ : str = [] for piece in pieces: if len(snake_case_ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): a_ : int = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case_ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: a_ : int = cur_pieces[1:] else: a_ : Tuple = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case_ ) else: new_pieces.append(snake_case_ ) return new_pieces def snake_case_ ( self , a_ ): return self.sp_model.PieceToId(snake_case_ ) def snake_case_ ( self , a_ ): return self.sp_model.IdToPiece(snake_case_ ) def snake_case_ ( self , a_ ): a_ : List[str] = "".join(snake_case_ ).replace(snake_case_ , " " ).strip() return out_string def snake_case_ ( self , a_ , a_ = False , a_ = None , a_ = True , **a_ , ): a_ : int = kwargs.pop("use_source_tokenizer" , snake_case_ ) a_ : Union[str, Any] = self.convert_ids_to_tokens(snake_case_ , skip_special_tokens=snake_case_ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 a_ : List[Any] = [] a_ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case_ ) ) a_ : List[str] = [] sub_texts.append(snake_case_ ) else: current_sub_text.append(snake_case_ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case_ ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens a_ : Tuple = "".join(snake_case_ ) a_ : Union[str, Any] = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: a_ : int = self.clean_up_tokenization(snake_case_ ) return clean_text else: return text def snake_case_ ( self , a_ , a_ = None ): a_ : Optional[int] = [self.sep_token_id] a_ : Tuple = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def snake_case_ ( self , a_ , a_ = None , a_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case_ , token_ids_a=snake_case_ , already_has_special_tokens=snake_case_ ) if token_ids_a is not None: return ([0] * len(snake_case_ )) + [1] + ([0] * len(snake_case_ )) + [1, 1] return ([0] * len(snake_case_ )) + [1, 1] def snake_case_ ( self , a_ , a_ = None ): a_ : Union[str, Any] = [self.sep_token_id] a_ : Optional[int] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def snake_case_ ( self , a_ , a_ = None ): if not os.path.isdir(snake_case_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return a_ : List[Any] = os.path.join( snake_case_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , "wb" ) as fi: a_ : str = self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
237
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase = { '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
104
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
"""simple docstring""" def a__ ( lowerCAmelCase ) -> List[str]: # noqa: E741 UpperCAmelCase__ : Optional[Any] = len(lowercase_ ) UpperCAmelCase__ : str = 0 UpperCAmelCase__ : List[Any] = [0] * n UpperCAmelCase__ : Tuple = [False] * n UpperCAmelCase__ : List[Any] = [False] * n def dfs(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if parent == root: out_edge_count += 1 UpperCAmelCase__ : Optional[int] = True UpperCAmelCase__ : Union[str, Any] = at for to in l[at]: if to == parent: pass elif not visited[to]: UpperCAmelCase__ : Any = dfs(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase__ : Union[str, Any] = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: UpperCAmelCase__ : Any = True # AP found via cycle if at == low[to]: UpperCAmelCase__ : Tuple = True else: UpperCAmelCase__ : int = min(low[at] , lowercase_ ) return out_edge_count for i in range(lowercase_ ): if not visited[i]: UpperCAmelCase__ : Union[str, Any] = 0 UpperCAmelCase__ : Tuple = dfs(lowercase_ , lowercase_ , -1 , lowercase_ ) UpperCAmelCase__ : List[Any] = out_edge_count > 1 for x in range(len(lowercase_ ) ): if is_art[x] is True: print(lowercase_ ) # Adjacency list of graph _A = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
182
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
0
import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=7, lowerCamelCase=6, lowerCamelCase=17, lowerCamelCase=23, lowerCamelCase=11, lowerCamelCase=True, ) -> Dict: """simple docstring""" _lowercase : Tuple = parent _lowercase : Dict = batch_size _lowercase : Dict = seq_length _lowercase : Tuple = act_dim _lowercase : int = state_dim _lowercase : List[Any] = hidden_size _lowercase : List[str] = max_length _lowercase : List[str] = is_training def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Tuple = floats_tensor((self.batch_size, self.seq_length, self.state_dim)) _lowercase : Union[str, Any] = floats_tensor((self.batch_size, self.seq_length, self.act_dim)) _lowercase : Dict = floats_tensor((self.batch_size, self.seq_length, 1)) _lowercase : Dict = floats_tensor((self.batch_size, self.seq_length, 1)) _lowercase : int = ids_tensor((self.batch_size, self.seq_length), vocab_size=10_00) _lowercase : str = random_attention_mask((self.batch_size, self.seq_length)) _lowercase : Dict = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def UpperCamelCase ( self) -> Any: """simple docstring""" return DecisionTransformerConfig( batch_size=self.batch_size, seq_length=self.seq_length, act_dim=self.act_dim, state_dim=self.state_dim, hidden_size=self.hidden_size, max_length=self.max_length, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[str]: """simple docstring""" _lowercase : List[str] = DecisionTransformerModel(config=snake_case_) model.to(snake_case_) model.eval() _lowercase : Any = model(snake_case_, snake_case_, snake_case_, snake_case_, snake_case_, snake_case_) self.parent.assertEqual(result.state_preds.shape, states.shape) self.parent.assertEqual(result.action_preds.shape, actions.shape) self.parent.assertEqual(result.return_preds.shape, returns_to_go.shape) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.seq_length * 3, self.hidden_size)) # seq length *3 as there are 3 modelities: states, returns and actions def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Dict = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Optional[Any] = config_and_inputs _lowercase : List[str] = { 'states': states, 'actions': actions, 'rewards': rewards, 'returns_to_go': returns_to_go, 'timesteps': timesteps, 'attention_mask': attention_mask, } return config, inputs_dict @require_torch class _lowerCamelCase( __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, unittest.TestCase ): lowercase_ : str = (DecisionTransformerModel,) if is_torch_available() else () lowercase_ : Any = () lowercase_ : List[str] = {"""feature-extraction""": DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids lowercase_ : Union[str, Any] = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features lowercase_ : List[Any] = False lowercase_ : Union[str, Any] = False lowercase_ : Optional[int] = False lowercase_ : Any = False lowercase_ : int = False lowercase_ : Optional[Any] = False lowercase_ : Optional[int] = False lowercase_ : Dict = False lowercase_ : List[Any] = False def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = DecisionTransformerModelTester(self) _lowercase : int = ConfigTester(self, config_class=snake_case_, hidden_size=37) def UpperCamelCase ( self) -> str: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_) @slow def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : List[str] = DecisionTransformerModel.from_pretrained(snake_case_) self.assertIsNotNone(snake_case_) def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase , _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : List[str] = model_class(snake_case_) _lowercase : int = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase : Dict = [*signature.parameters.keys()] _lowercase : Union[str, Any] = [ 'states', 'actions', 'rewards', 'returns_to_go', 'timesteps', 'attention_mask', ] self.assertListEqual(arg_names[: len(snake_case_)], snake_case_) @require_torch class _lowerCamelCase( unittest.TestCase ): @slow def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Any = 2 # number of steps of autoregressive prediction we will perform _lowercase : List[str] = 10 # defined by the RL environment, may be normalized _lowercase : int = DecisionTransformerModel.from_pretrained('edbeeching/decision-transformer-gym-hopper-expert') _lowercase : Optional[Any] = model.to(snake_case_) _lowercase : Optional[Any] = model.config torch.manual_seed(0) _lowercase : int = torch.randn(1, 1, config.state_dim).to(device=snake_case_, dtype=torch.floataa) # env.reset() _lowercase : List[Any] = torch.tensor( [[0.2_4_2_7_9_3, -0.2_8_6_9_3_0_7_4, 0.8_7_4_2_6_1_3], [0.6_7_8_1_5_2_7_4, -0.0_8_1_0_1_0_8_5, -0.1_2_9_5_2_1_4_7]], device=snake_case_) _lowercase : Optional[Any] = torch.tensor(snake_case_, device=snake_case_, dtype=torch.floataa).reshape(1, 1, 1) _lowercase : Union[str, Any] = state _lowercase : List[str] = torch.zeros(1, 0, config.act_dim, device=snake_case_, dtype=torch.floataa) _lowercase : int = torch.zeros(1, 0, device=snake_case_, dtype=torch.floataa) _lowercase : List[str] = torch.tensor(0, device=snake_case_, dtype=torch.long).reshape(1, 1) for step in range(snake_case_): _lowercase : str = torch.cat([actions, torch.zeros(1, 1, config.act_dim, device=snake_case_)], dim=1) _lowercase : Tuple = torch.cat([rewards, torch.zeros(1, 1, device=snake_case_)], dim=1) _lowercase : Dict = torch.ones(1, states.shape[1]).to(dtype=torch.long, device=states.device) with torch.no_grad(): _lowercase , _lowercase , _lowercase : Optional[Any] = model( states=snake_case_, actions=snake_case_, rewards=snake_case_, returns_to_go=snake_case_, timesteps=snake_case_, attention_mask=snake_case_, return_dict=snake_case_, ) self.assertEqual(action_pred.shape, actions.shape) self.assertTrue(torch.allclose(action_pred[0, -1], expected_outputs[step], atol=1E-4)) _lowercase , _lowercase , _lowercase , _lowercase : str = ( # env.step(action) torch.randn(1, 1, config.state_dim).to(device=snake_case_, dtype=torch.floataa), 1.0, False, {}, ) _lowercase : str = action_pred[0, -1] _lowercase : Tuple = torch.cat([states, state], dim=1) _lowercase : List[str] = returns_to_go[0, -1] - reward _lowercase : int = torch.cat([returns_to_go, pred_return.reshape(1, 1, 1)], dim=1) _lowercase : int = torch.cat( [timesteps, torch.ones((1, 1), device=snake_case_, dtype=torch.long) * (step + 1)], dim=1)
89
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCamelCase = { '''configuration_blenderbot''': [ '''BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlenderbotConfig''', '''BlenderbotOnnxConfig''', ], '''tokenization_blenderbot''': ['''BlenderbotTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['''BlenderbotTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ '''BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlenderbotForCausalLM''', '''BlenderbotForConditionalGeneration''', '''BlenderbotModel''', '''BlenderbotPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ '''TFBlenderbotForConditionalGeneration''', '''TFBlenderbotModel''', '''TFBlenderbotPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ '''FlaxBlenderbotForConditionalGeneration''', '''FlaxBlenderbotModel''', '''FlaxBlenderbotPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
114
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
0
import math def UpperCamelCase ( ): '''simple docstring''' A_ : Optional[Any] = input('Enter message: ' ) A_ : List[Any] = int(input(f'''Enter key [2-{len(lowercase_ ) - 1}]: ''' ) ) A_ : Tuple = input('Encryption/Decryption [e/d]: ' ) if mode.lower().startswith('e' ): A_ : List[str] = encrypt_message(lowercase_ ,lowercase_ ) elif mode.lower().startswith('d' ): A_ : Any = decrypt_message(lowercase_ ,lowercase_ ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f'''Output:\n{text + '|'}''' ) def UpperCamelCase ( __lowercase : int ,__lowercase : str ): '''simple docstring''' A_ : Optional[int] = [''] * key for col in range(lowercase_ ): A_ : str = col while pointer < len(lowercase_ ): cipher_text[col] += message[pointer] pointer += key return "".join(lowercase_ ) def UpperCamelCase ( __lowercase : int ,__lowercase : str ): '''simple docstring''' A_ : int = math.ceil(len(lowercase_ ) / key ) A_ : Any = key A_ : Optional[int] = (num_cols * num_rows) - len(lowercase_ ) A_ : str = [''] * num_cols A_ : List[str] = 0 A_ : Any = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): A_ : List[str] = 0 row += 1 return "".join(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
558
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
0
'''simple docstring''' from __future__ import annotations import math def _UpperCAmelCase ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : bool , _lowerCamelCase : list[int] , _lowerCamelCase : float ) -> int: if depth < 0: raise ValueError("""Depth cannot be less than 0""" ) if len(lowercase_ ) == 0: raise ValueError("""Scores cannot be empty""" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 , node_index * 2 , lowercase_ , lowercase_ , lowercase_ ) , minimax(depth + 1 , node_index * 2 + 1 , lowercase_ , lowercase_ , lowercase_ ) , ) return min( minimax(depth + 1 , node_index * 2 , lowercase_ , lowercase_ , lowercase_ ) , minimax(depth + 1 , node_index * 2 + 1 , lowercase_ , lowercase_ , lowercase_ ) , ) def _UpperCAmelCase ( ) -> None: _lowerCAmelCase : Optional[Any] = [90, 23, 6, 33, 21, 65, 1_23, 3_44_23] _lowerCAmelCase : int = math.log(len(lowercase_ ) , 2 ) print("""Optimal value : """ , end="""""" ) print(minimax(0 , 0 , lowercase_ , lowercase_ , lowercase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
384
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
0
"""simple docstring""" from string import ascii_uppercase UpperCamelCase__ = {str(ord(c) - 5_5): c for c in ascii_uppercase} def UpperCAmelCase ( snake_case : int , snake_case : int ): if isinstance(lowercase_ , lowercase_ ): raise TypeError('''int() can\'t convert non-string with explicit base''' ) if num < 0: raise ValueError('''parameter must be positive int''' ) if isinstance(lowercase_ , lowercase_ ): raise TypeError('''\'str\' object cannot be interpreted as an integer''' ) if isinstance(lowercase_ , lowercase_ ): raise TypeError('''\'float\' object cannot be interpreted as an integer''' ) if base in (0, 1): raise ValueError('''base must be >= 2''' ) if base > 36: raise ValueError('''base must be <= 36''' ) _lowerCAmelCase:int = '''''' _lowerCAmelCase:Tuple = 0 _lowerCAmelCase:Optional[int] = 0 while div != 1: _lowerCAmelCase , _lowerCAmelCase:Any = divmod(lowercase_ , lowercase_ ) if base >= 11 and 9 < mod < 36: _lowerCAmelCase:Optional[int] = ALPHABET_VALUES[str(lowercase_ )] else: _lowerCAmelCase:Tuple = str(lowercase_ ) new_value += actual_value _lowerCAmelCase:Optional[Any] = num // base _lowerCAmelCase:List[str] = div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(lowercase_ ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 3_7): for num in range(1_0_0_0): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
227
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
280
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
0
'''simple docstring''' import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def __init__(self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=False , lowercase=True , lowercase="None" , lowercase=3 , lowercase=4 , lowercase=None , ): A_ : Tuple = parent A_ : Tuple = batch_size A_ : Any = seq_length A_ : Dict = is_training A_ : Optional[int] = use_input_mask A_ : List[str] = use_token_type_ids A_ : str = use_labels A_ : Dict = vocab_size A_ : List[str] = hidden_size A_ : Any = num_hidden_layers A_ : Optional[Any] = num_attention_heads A_ : Dict = intermediate_size A_ : str = hidden_act A_ : Any = hidden_dropout_prob A_ : Tuple = attention_probs_dropout_prob A_ : Tuple = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : Optional[Any] = type_sequence_label_size A_ : Tuple = initializer_range A_ : Dict = num_labels A_ : Tuple = num_choices A_ : List[str] = relative_attention A_ : Optional[int] = position_biased_input A_ : Dict = pos_att_type A_ : Optional[int] = scope def _a (self ): A_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : Dict = None if self.use_input_mask: A_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) A_ : Any = None if self.use_token_type_ids: A_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : Tuple = None A_ : List[str] = None A_ : Tuple = None if self.use_labels: A_ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) A_ : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _a (self ): return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def _a (self ): A_ : List[str] = self.get_config() A_ : int = 300 return config def _a (self , lowercase ): self.parent.assertListEqual(list(result.loss.size() ) , [] ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Optional[Any] = DebertaModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() A_ : List[str] = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ )[0] A_ : str = model(snake_case_ , token_type_ids=snake_case_ )[0] A_ : str = model(snake_case_ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : str = DebertaForMaskedLM(config=snake_case_ ) model.to(snake_case_ ) model.eval() A_ : Optional[Any] = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : int = self.num_labels A_ : List[str] = DebertaForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() A_ : Any = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(snake_case_ ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : List[Any] = self.num_labels A_ : Optional[int] = DebertaForTokenClassification(config=snake_case_ ) model.to(snake_case_ ) model.eval() A_ : List[Any] = model(snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a (self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Optional[Any] = DebertaForQuestionAnswering(config=snake_case_ ) model.to(snake_case_ ) model.eval() A_ : List[str] = model( snake_case_ , attention_mask=snake_case_ , token_type_ids=snake_case_ , start_positions=snake_case_ , end_positions=snake_case_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a (self ): A_ : List[str] = self.prepare_config_and_inputs() ( ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ( A_ ), ) : Optional[int] = config_and_inputs A_ : Dict = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): __SCREAMING_SNAKE_CASE : str = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': DebertaModel, 'fill-mask': DebertaForMaskedLM, 'question-answering': DebertaForQuestionAnswering, 'text-classification': DebertaForSequenceClassification, 'token-classification': DebertaForTokenClassification, 'zero-shot': DebertaForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE : Any = True __SCREAMING_SNAKE_CASE : List[Any] = False __SCREAMING_SNAKE_CASE : Union[str, Any] = False __SCREAMING_SNAKE_CASE : Optional[int] = False __SCREAMING_SNAKE_CASE : Dict = False def _a (self ): A_ : List[Any] = DebertaModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _a (self ): self.config_tester.run_common_tests() def _a (self ): A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*snake_case_ ) def _a (self ): A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*snake_case_ ) def _a (self ): A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*snake_case_ ) def _a (self ): A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*snake_case_ ) def _a (self ): A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*snake_case_ ) @slow def _a (self ): for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Any = DebertaModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @require_torch @require_sentencepiece @require_tokenizers class _lowerCAmelCase ( unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def _a (self ): pass @slow def _a (self ): A_ : Tuple = DebertaModel.from_pretrained("""microsoft/deberta-base""" ) A_ : int = torch.tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) A_ : List[str] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): A_ : Optional[Any] = model(snake_case_ , attention_mask=snake_case_ )[0] # compare the actual values for a slice. A_ : int = torch.tensor( [[[-0.59_86, -0.80_55, -0.84_62], [1.44_84, -0.93_48, -0.80_59], [0.31_23, 0.00_32, -1.41_31]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , snake_case_ , atol=1E-4 ) , F'{output[:, 1:4, 1:4]}' )
667
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[int] = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase : List[str] ="vit_msn" def __init__( self : Dict , lowerCAmelCase : Dict=7_68 , lowerCAmelCase : Any=12 , lowerCAmelCase : Dict=12 , lowerCAmelCase : Dict=30_72 , lowerCAmelCase : Optional[Any]="gelu" , lowerCAmelCase : str=0.0 , lowerCAmelCase : Optional[Any]=0.0 , lowerCAmelCase : Optional[Any]=0.0_2 , lowerCAmelCase : str=1e-06 , lowerCAmelCase : Tuple=2_24 , lowerCAmelCase : List[Any]=16 , lowerCAmelCase : Union[str, Any]=3 , lowerCAmelCase : int=True , **lowerCAmelCase : Any , ): super().__init__(**snake_case_ ) A_ = hidden_size A_ = num_hidden_layers A_ = num_attention_heads A_ = intermediate_size A_ = hidden_act A_ = hidden_dropout_prob A_ = attention_probs_dropout_prob A_ = initializer_range A_ = layer_norm_eps A_ = image_size A_ = patch_size A_ = num_channels A_ = qkv_bias
452
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ ) -> int: a_ : Optional[Any] = args.pruning_method a_ : Tuple = args.threshold a_ : int = args.model_name_or_path.rstrip("/" ) a_ : Optional[Any] = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) a_ : Union[str, Any] = torch.load(os.path.join(lowercase_, "pytorch_model.bin" ) ) a_ : Any = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: a_ : Optional[int] = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: a_ : Any = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: a_ : List[Any] = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": a_ : Union[str, Any] = MagnitudeBinarizer.apply(inputs=lowercase_, threshold=lowercase_ ) a_ : Any = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue a_ : Optional[Any] = name[:-6] a_ : str = model[F"""{prefix_}mask_scores"""] a_ : Optional[int] = TopKBinarizer.apply(lowercase_, lowercase_ ) a_ : Optional[Any] = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue a_ : Tuple = name[:-6] a_ : Any = model[F"""{prefix_}mask_scores"""] a_ : List[str] = ThresholdBinarizer.apply(lowercase_, lowercase_, lowercase_ ) a_ : Tuple = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue a_ : str = name[:-6] a_ : Union[str, Any] = model[F"""{prefix_}mask_scores"""] a_ , a_ : Optional[Any] = -0.1, 1.1 a_ : Union[str, Any] = torch.sigmoid(lowercase_ ) a_ : Tuple = s * (r - l) + l a_ : List[str] = s_bar.clamp(min=0.0, max=1.0 ) a_ : List[str] = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError("Unknown pruning method" ) if target_model_path is None: a_ : Optional[int] = os.path.join( os.path.dirname(lowercase_ ), F"""bertarized_{os.path.basename(lowercase_ )}""" ) if not os.path.isdir(lowercase_ ): shutil.copytree(lowercase_, lowercase_ ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(lowercase_, os.path.join(lowercase_, "pytorch_model.bin" ) ) print("\nPruned model saved! See you later!" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) SCREAMING_SNAKE_CASE_ = parser.parse_args() main(args)
237
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
0
"""simple docstring""" import gc import math import unittest import torch from diffusers import UNetaDModel from diffusers.utils import floats_tensor, logging, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin UpperCamelCase = logging.get_logger(__name__) enable_full_determinism() class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" A__ : Any = UNetaDModel A__ : int = "sample" @property def snake_case__ ( self ) -> Optional[int]: A__ = 4 A__ = 3 A__ = (32, 32) A__ = floats_tensor((batch_size, num_channels) + sizes ).to(snake_case_ ) A__ = torch.tensor([10] ).to(snake_case_ ) return {"sample": noise, "timestep": time_step} @property def snake_case__ ( self ) -> Union[str, Any]: return (3, 32, 32) @property def snake_case__ ( self ) -> Union[str, Any]: return (3, 32, 32) def snake_case__ ( self ) -> Optional[Any]: A__ = { "block_out_channels": (32, 64), "down_block_types": ("DownBlock2D", "AttnDownBlock2D"), "up_block_types": ("AttnUpBlock2D", "UpBlock2D"), "attention_head_dim": 3, "out_channels": 3, "in_channels": 3, "layers_per_block": 2, "sample_size": 32, } A__ = self.dummy_input return init_dict, inputs_dict class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" A__ : Union[str, Any] = UNetaDModel A__ : int = "sample" @property def snake_case__ ( self ) -> Dict: A__ = 4 A__ = 4 A__ = (32, 32) A__ = floats_tensor((batch_size, num_channels) + sizes ).to(snake_case_ ) A__ = torch.tensor([10] ).to(snake_case_ ) return {"sample": noise, "timestep": time_step} @property def snake_case__ ( self ) -> Union[str, Any]: return (4, 32, 32) @property def snake_case__ ( self ) -> Tuple: return (4, 32, 32) def snake_case__ ( self ) -> str: A__ = { "sample_size": 32, "in_channels": 4, "out_channels": 4, "layers_per_block": 2, "block_out_channels": (32, 64), "attention_head_dim": 32, "down_block_types": ("DownBlock2D", "DownBlock2D"), "up_block_types": ("UpBlock2D", "UpBlock2D"), } A__ = self.dummy_input return init_dict, inputs_dict def snake_case__ ( self ) -> Union[str, Any]: A__ , A__ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=snake_case_ ) self.assertIsNotNone(snake_case_ ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(snake_case_ ) A__ = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != "cuda" , "This test is supposed to run on GPU" ) def snake_case__ ( self ) -> int: A__ , A__ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=snake_case_ ) model.to(snake_case_ ) A__ = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != "cuda" , "This test is supposed to run on GPU" ) def snake_case__ ( self ) -> Dict: # by defautl model loading will use accelerate as `low_cpu_mem_usage=True` A__ , A__ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" , output_loading_info=snake_case_ ) model_accelerate.to(snake_case_ ) model_accelerate.eval() A__ = torch.randn( 1 , model_accelerate.config.in_channels , model_accelerate.config.sample_size , model_accelerate.config.sample_size , generator=torch.manual_seed(0 ) , ) A__ = noise.to(snake_case_ ) A__ = torch.tensor([10] * noise.shape[0] ).to(snake_case_ ) A__ = model_accelerate(snake_case_ , snake_case_ )["sample"] # two models don't need to stay in the device at the same time del model_accelerate torch.cuda.empty_cache() gc.collect() A__ , A__ = UNetaDModel.from_pretrained( "fusing/unet-ldm-dummy-update" , output_loading_info=snake_case_ , low_cpu_mem_usage=snake_case_ ) model_normal_load.to(snake_case_ ) model_normal_load.eval() A__ = model_normal_load(snake_case_ , snake_case_ )["sample"] assert torch_all_close(snake_case_ , snake_case_ , rtol=1e-3 ) def snake_case__ ( self ) -> int: A__ = UNetaDModel.from_pretrained("fusing/unet-ldm-dummy-update" ) model.eval() model.to(snake_case_ ) A__ = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) A__ = noise.to(snake_case_ ) A__ = torch.tensor([10] * noise.shape[0] ).to(snake_case_ ) with torch.no_grad(): A__ = model(snake_case_ , snake_case_ ).sample A__ = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off A__ = torch.tensor([-1_3.3_2_5_8, -2_0.1_1_0_0, -1_5.9_8_7_3, -1_7.6_6_1_7, -2_3.0_5_9_6, -1_7.9_4_1_9, -1_3.3_6_7_5, -1_6.1_8_8_9, -1_2.3_8_0_0] ) # fmt: on self.assertTrue(torch_all_close(snake_case_ , snake_case_ , rtol=1e-3 ) ) class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" A__ : Optional[Any] = UNetaDModel A__ : Dict = "sample" @property def snake_case__ ( self , SCREAMING_SNAKE_CASE__=(32, 32) ) -> Optional[Any]: A__ = 4 A__ = 3 A__ = floats_tensor((batch_size, num_channels) + sizes ).to(snake_case_ ) A__ = torch.tensor(batch_size * [10] ).to(dtype=torch.intaa , device=snake_case_ ) return {"sample": noise, "timestep": time_step} @property def snake_case__ ( self ) -> int: return (3, 32, 32) @property def snake_case__ ( self ) -> Any: return (3, 32, 32) def snake_case__ ( self ) -> int: A__ = { "block_out_channels": [32, 64, 64, 64], "in_channels": 3, "layers_per_block": 1, "out_channels": 3, "time_embedding_type": "fourier", "norm_eps": 1e-6, "mid_block_scale_factor": math.sqrt(2.0 ), "norm_num_groups": None, "down_block_types": [ "SkipDownBlock2D", "AttnSkipDownBlock2D", "SkipDownBlock2D", "SkipDownBlock2D", ], "up_block_types": [ "SkipUpBlock2D", "SkipUpBlock2D", "AttnSkipUpBlock2D", "SkipUpBlock2D", ], } A__ = self.dummy_input return init_dict, inputs_dict @slow def snake_case__ ( self ) -> Tuple: A__ , A__ = UNetaDModel.from_pretrained("google/ncsnpp-celebahq-256" , output_loading_info=snake_case_ ) self.assertIsNotNone(snake_case_ ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(snake_case_ ) A__ = self.dummy_input A__ = floats_tensor((4, 3) + (256, 256) ).to(snake_case_ ) A__ = noise A__ = model(**snake_case_ ) assert image is not None, "Make sure output is not None" @slow def snake_case__ ( self ) -> List[str]: A__ = UNetaDModel.from_pretrained("google/ncsnpp-celebahq-256" ) model.to(snake_case_ ) A__ = 4 A__ = 3 A__ = (256, 256) A__ = torch.ones((batch_size, num_channels) + sizes ).to(snake_case_ ) A__ = torch.tensor(batch_size * [1e-4] ).to(snake_case_ ) with torch.no_grad(): A__ = model(snake_case_ , snake_case_ ).sample A__ = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off A__ = torch.tensor([-4842.8691, -6499.6631, -3800.1953, -7978.2686, -10980.7129, -20028.8535, 8148.2822, 2342.2905, 567.7608] ) # fmt: on self.assertTrue(torch_all_close(snake_case_ , snake_case_ , rtol=1e-2 ) ) def snake_case__ ( self ) -> Any: A__ = UNetaDModel.from_pretrained("fusing/ncsnpp-ffhq-ve-dummy-update" ) model.to(snake_case_ ) A__ = 4 A__ = 3 A__ = (32, 32) A__ = torch.ones((batch_size, num_channels) + sizes ).to(snake_case_ ) A__ = torch.tensor(batch_size * [1e-4] ).to(snake_case_ ) with torch.no_grad(): A__ = model(snake_case_ , snake_case_ ).sample A__ = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off A__ = torch.tensor([-0.0_3_2_5, -0.0_9_0_0, -0.0_8_6_9, -0.0_3_3_2, -0.0_7_2_5, -0.0_2_7_0, -0.0_1_0_1, 0.0_2_2_7, 0.0_2_5_6] ) # fmt: on self.assertTrue(torch_all_close(snake_case_ , snake_case_ , rtol=1e-2 ) ) def snake_case__ ( self ) -> List[str]: # not required for this model pass
104
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A = { '''configuration_xlm_roberta_xl''': [ '''XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMRobertaXLConfig''', '''XLMRobertaXLOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMRobertaXLForCausalLM''', '''XLMRobertaXLForMaskedLM''', '''XLMRobertaXLForMultipleChoice''', '''XLMRobertaXLForQuestionAnswering''', '''XLMRobertaXLForSequenceClassification''', '''XLMRobertaXLForTokenClassification''', '''XLMRobertaXLModel''', '''XLMRobertaXLPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
182
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
0
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py SCREAMING_SNAKE_CASE : Any = '''src/diffusers''' SCREAMING_SNAKE_CASE : str = '''.''' # This is to make sure the diffusers module imported is the one in the repo. SCREAMING_SNAKE_CASE : List[str] = importlib.util.spec_from_file_location( "diffusers", os.path.join(DIFFUSERS_PATH, "__init__.py"), submodule_search_locations=[DIFFUSERS_PATH], ) SCREAMING_SNAKE_CASE : Optional[Any] = spec.loader.load_module() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> List[Any]: return line.startswith(lowercase_ ) or len(lowercase_ ) <= 1 or re.search(R'^\s*\)(\s*->.*:|:)\s*$' , lowercase_ ) is not None def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Optional[int] = object_name.split('.' ) _lowercase : List[str] = 0 # First let's find the module where our object lives. _lowercase : Union[str, Any] = parts[i] while i < len(lowercase_ ) and not os.path.isfile(os.path.join(lowercase_ , F'''{module}.py''' ) ): i += 1 if i < len(lowercase_ ): _lowercase : Any = os.path.join(lowercase_ , parts[i] ) if i >= len(lowercase_ ): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''' ) with open(os.path.join(lowercase_ , F'''{module}.py''' ) , 'r' , encoding='utf-8' , newline='\n' ) as f: _lowercase : Union[str, Any] = f.readlines() # Now let's find the class / func in the code! _lowercase : Any = '' _lowercase : Any = 0 for name in parts[i + 1 :]: while ( line_index < len(lowercase_ ) and re.search(RF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowercase_ ): raise ValueError(F''' {object_name} does not match any function or class in {module}.''' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). _lowercase : str = line_index while line_index < len(lowercase_ ) and _should_continue(lines[line_index] , lowercase_ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 _lowercase : int = lines[start_index:line_index] return "".join(lowercase_ ) SCREAMING_SNAKE_CASE : Any = re.compile(r"^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)") SCREAMING_SNAKE_CASE : Optional[int] = re.compile(r"^\s*(\S+)->(\S+)(\s+.*|$)") SCREAMING_SNAKE_CASE : Optional[int] = re.compile(r"<FILL\s+[^>]*>") def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: _lowercase : int = code.split('\n' ) _lowercase : List[str] = 0 while idx < len(lowercase_ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowercase_ ): return re.search(R'^(\s*)\S' , lines[idx] ).groups()[0] return "" def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: _lowercase : Union[str, Any] = len(get_indent(lowercase_ ) ) > 0 if has_indent: _lowercase : Optional[Any] = F'''class Bla:\n{code}''' _lowercase : Any = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=lowercase_ ) _lowercase : Union[str, Any] = black.format_str(lowercase_ , mode=lowercase_ ) _lowercase , _lowercase : str = style_docstrings_in_code(lowercase_ ) return result[len('class Bla:\n' ) :] if has_indent else result def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> int: with open(lowercase_ , 'r' , encoding='utf-8' , newline='\n' ) as f: _lowercase : List[Any] = f.readlines() _lowercase : Union[str, Any] = [] _lowercase : Any = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowercase_ ): _lowercase : Tuple = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. _lowercase , _lowercase , _lowercase : int = search.groups() _lowercase : List[Any] = find_code_in_diffusers(lowercase_ ) _lowercase : Optional[Any] = get_indent(lowercase_ ) _lowercase : Optional[Any] = line_index + 1 if indent == theoretical_indent else line_index + 2 _lowercase : int = theoretical_indent _lowercase : List[Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. _lowercase : Union[str, Any] = True while line_index < len(lowercase_ ) and should_continue: line_index += 1 if line_index >= len(lowercase_ ): break _lowercase : Union[str, Any] = lines[line_index] _lowercase : str = _should_continue(lowercase_ , lowercase_ ) and re.search(F'''^{indent}# End copy''' , lowercase_ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 _lowercase : Tuple = lines[start_index:line_index] _lowercase : Tuple = ''.join(lowercase_ ) # Remove any nested `Copied from` comments to avoid circular copies _lowercase : Optional[Any] = [line for line in theoretical_code.split('\n' ) if _re_copy_warning.search(lowercase_ ) is None] _lowercase : Dict = '\n'.join(lowercase_ ) # Before comparing, use the `replace_pattern` on the original code. if len(lowercase_ ) > 0: _lowercase : List[str] = replace_pattern.replace('with' , '' ).split(',' ) _lowercase : List[Any] = [_re_replace_pattern.search(lowercase_ ) for p in patterns] for pattern in patterns: if pattern is None: continue _lowercase , _lowercase , _lowercase : int = pattern.groups() _lowercase : Tuple = re.sub(lowercase_ , lowercase_ , lowercase_ ) if option.strip() == "all-casing": _lowercase : Dict = re.sub(obja.lower() , obja.lower() , lowercase_ ) _lowercase : Tuple = re.sub(obja.upper() , obja.upper() , lowercase_ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line _lowercase : List[str] = blackify(lines[start_index - 1] + theoretical_code ) _lowercase : Any = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: _lowercase : Tuple = lines[:start_index] + [theoretical_code] + lines[line_index:] _lowercase : List[str] = start_index + 1 if overwrite and len(lowercase_ ) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''' ) with open(lowercase_ , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(lowercase_ ) return diffs def UpperCamelCase_( lowerCamelCase_ = False ) -> Dict: _lowercase : Union[str, Any] = glob.glob(os.path.join(lowercase_ , '**/*.py' ) , recursive=lowercase_ ) _lowercase : int = [] for filename in all_files: _lowercase : Optional[Any] = is_copy_consistent(lowercase_ , lowercase_ ) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(lowercase_ ) > 0: _lowercase : Any = '\n'.join(lowercase_ ) raise Exception( 'Found the following copy inconsistencies:\n' + diff + '\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") SCREAMING_SNAKE_CASE : Dict = parser.parse_args() check_copies(args.fix_and_overwrite)
89
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
0
def __UpperCAmelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_=None ): _lowerCamelCase : int = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: _lowerCamelCase, _lowerCamelCase : List[str] = True, True _lowerCamelCase : Union[str, Any] = dfs(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) return path def __UpperCAmelCase( lowercase_ , lowercase_ ): _lowerCamelCase : Dict = 0 _lowerCamelCase : str = -1 for i in range(lowercase_ ): if i not in graph.keys(): continue if len(graph[i] ) % 2 == 1: odd_degree_nodes += 1 _lowerCamelCase : Dict = i if odd_degree_nodes == 0: return 1, odd_node if odd_degree_nodes == 2: return 2, odd_node return 3, odd_node def __UpperCAmelCase( lowercase_ , lowercase_ ): _lowerCamelCase : Optional[int] = [[False for _ in range(max_node + 1 )] for _ in range(max_node + 1 )] _lowerCamelCase, _lowerCamelCase : str = check_circuit_or_path(lowercase_ , lowercase_ ) if check == 3: print('''graph is not Eulerian''' ) print('''no path''' ) return _lowerCamelCase : List[str] = 1 if check == 2: _lowerCamelCase : str = odd_node print('''graph has a Euler path''' ) if check == 1: print('''graph has a Euler cycle''' ) _lowerCamelCase : List[Any] = dfs(lowercase_ , lowercase_ , lowercase_ ) print(lowercase_ ) def __UpperCAmelCase( ): _lowerCamelCase : Optional[Any] = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]} _lowerCamelCase : Union[str, Any] = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]} _lowerCamelCase : Union[str, Any] = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]} _lowerCamelCase : List[Any] = {1: [2, 3], 2: [1, 3], 3: [1, 2]} _lowerCamelCase : Dict = { 1: [], 2: [] # all degree is zero } _lowerCamelCase : Tuple = 10 check_euler(lowercase_ , lowercase_ ) check_euler(lowercase_ , lowercase_ ) check_euler(lowercase_ , lowercase_ ) check_euler(lowercase_ , lowercase_ ) check_euler(lowercase_ , lowercase_ ) if __name__ == "__main__": main()
114
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
0
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase = version.parse(importlib_metadata.version("""nltk""")) if NLTK_VERSION >= version.Version("""3.6.4"""): from nltk import word_tokenize _UpperCAmelCase = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def lowerCAmelCase_ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'] , reference_urls=[ 'https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score', 'https://en.wikipedia.org/wiki/METEOR', ] , ) def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" import nltk nltk.download('wordnet' ) if NLTK_VERSION >= version.Version('3.6.5' ): nltk.download('punkt' ) if NLTK_VERSION >= version.Version('3.6.6' ): nltk.download('omw-1.4' ) def lowerCAmelCase_ ( self , lowercase , lowercase , lowercase=0.9 , lowercase=3 , lowercase=0.5 ): """simple docstring""" if NLTK_VERSION >= version.Version('3.6.5' ): A_ : Optional[Any] = [ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: A_ : Optional[Any] = [ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
558
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
0
'''simple docstring''' import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def _UpperCAmelCase ( _lowerCamelCase : List[str] , _lowerCamelCase : Union[str, Any]=10 ) -> List[Any]: _lowerCAmelCase : int = [] for _ in range(lowercase_ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def _UpperCAmelCase ( _lowerCamelCase : List[Any] , _lowerCamelCase : str=10 ) -> Dict: _lowerCAmelCase : Tuple = [] for step in range(lowercase_ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: _lowerCAmelCase : Optional[Any] = os.path.join(lowercase_ , """schedule.bin""" ) torch.save(scheduler.state_dict() , lowercase_ ) _lowerCAmelCase : Dict = torch.load(lowercase_ ) scheduler.load_state_dict(lowercase_ ) return lrs @require_torch class a_ (unittest.TestCase ): def __UpperCamelCase ( self , snake_case_ , snake_case_ , snake_case_ ): self.assertEqual(len(snake_case_ ) , len(snake_case_ ) ) for a, b in zip(snake_case_ , snake_case_ ): self.assertAlmostEqual(snake_case_ , snake_case_ , delta=snake_case_ ) def __UpperCamelCase ( self ): _lowerCAmelCase : Dict = torch.tensor([0.1, -0.2, -0.1] , requires_grad=snake_case_ ) _lowerCAmelCase : Union[str, Any] = torch.tensor([0.4, 0.2, -0.5] ) _lowerCAmelCase : Optional[int] = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _lowerCAmelCase : List[Any] = AdamW(params=[w] , lr=2E-1 , weight_decay=0.0 ) for _ in range(1_0_0 ): _lowerCAmelCase : Union[str, Any] = criterion(snake_case_ , snake_case_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) def __UpperCamelCase ( self ): _lowerCAmelCase : List[Any] = torch.tensor([0.1, -0.2, -0.1] , requires_grad=snake_case_ ) _lowerCAmelCase : Any = torch.tensor([0.4, 0.2, -0.5] ) _lowerCAmelCase : List[Any] = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _lowerCAmelCase : Union[str, Any] = Adafactor( params=[w] , lr=1E-2 , eps=(1E-30, 1E-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=snake_case_ , weight_decay=0.0 , relative_step=snake_case_ , scale_parameter=snake_case_ , warmup_init=snake_case_ , ) for _ in range(1_0_0_0 ): _lowerCAmelCase : int = criterion(snake_case_ , snake_case_ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1E-2 ) @require_torch class a_ (unittest.TestCase ): __lowerCAmelCase : List[Any] = nn.Linear(5_0 , 5_0 ) if is_torch_available() else None __lowerCAmelCase : str = AdamW(m.parameters() , lr=1_0.0 ) if is_torch_available() else None __lowerCAmelCase : List[str] = 1_0 def __UpperCamelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): self.assertEqual(len(snake_case_ ) , len(snake_case_ ) ) for a, b in zip(snake_case_ , snake_case_ ): self.assertAlmostEqual(snake_case_ , snake_case_ , delta=snake_case_ , msg=snake_case_ ) def __UpperCamelCase ( self ): _lowerCAmelCase : int = {"""num_warmup_steps""": 2, """num_training_steps""": 1_0} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) _lowerCAmelCase : List[str] = { get_constant_schedule: ({}, [10.0] * self.num_steps), get_constant_schedule_with_warmup: ( {"""num_warmup_steps""": 4}, [0.0, 2.5, 5.0, 7.5, 10.0, 10.0, 10.0, 10.0, 10.0, 10.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 8.75, 7.5, 6.25, 5.0, 3.75, 2.5, 1.25], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 10.0, 9.61, 8.53, 6.91, 5.0, 3.08, 1.46, 0.38], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, """num_cycles""": 2}, [0.0, 5.0, 10.0, 8.53, 5.0, 1.46, 10.0, 8.53, 5.0, 1.46], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, """power""": 2.0, """lr_end""": 1E-7}, [0.0, 5.0, 10.0, 7.656, 5.625, 3.906, 2.5, 1.406, 0.625, 0.156], ), get_inverse_sqrt_schedule: ( {"""num_warmup_steps""": 2}, [0.0, 5.0, 10.0, 8.165, 7.071, 6.325, 5.774, 5.345, 5.0, 4.714], ), } for scheduler_func, data in scheds.items(): _lowerCAmelCase , _lowerCAmelCase : int = data _lowerCAmelCase : Tuple = scheduler_func(self.optimizer , **snake_case_ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) _lowerCAmelCase : Union[str, Any] = unwrap_schedule(snake_case_ , self.num_steps ) self.assertListAlmostEqual( snake_case_ , snake_case_ , tol=1E-2 , msg=f'failed for {scheduler_func} in normal scheduler' , ) _lowerCAmelCase : Optional[int] = scheduler_func(self.optimizer , **snake_case_ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(snake_case_ ) # wrap to test picklability of the schedule _lowerCAmelCase : Union[str, Any] = unwrap_and_save_reload_schedule(snake_case_ , self.num_steps ) self.assertListEqual(snake_case_ , snake_case_ , msg=f'failed for {scheduler_func} in save and reload' ) class a_ : def __init__( self , snake_case_ ): _lowerCAmelCase : List[Any] = fn def __call__( self , *snake_case_ , **snake_case_ ): return self.fn(*snake_case_ , **snake_case_ ) @classmethod def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : Any = list(map(self , scheduler.lr_lambdas ) )
384
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
0
"""simple docstring""" import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def UpperCAmelCase ( snake_case : Tuple ): return 1 / (1 + np.exp(-z )) def UpperCAmelCase ( snake_case : Optional[int] , snake_case : Dict ): return (-y * np.log(lowercase_ ) - (1 - y) * np.log(1 - h )).mean() def UpperCAmelCase ( snake_case : Optional[int] , snake_case : Tuple , snake_case : Tuple ): _lowerCAmelCase:Dict = np.dot(lowercase_ , lowercase_ ) return np.sum(y * scores - np.log(1 + np.exp(lowercase_ ) ) ) def UpperCAmelCase ( snake_case : int , snake_case : Tuple , snake_case : str , snake_case : Optional[int]=70000 ): _lowerCAmelCase:Dict = np.zeros(x.shape[1] ) for iterations in range(lowercase_ ): _lowerCAmelCase:List[Any] = np.dot(lowercase_ , lowercase_ ) _lowerCAmelCase:str = sigmoid_function(lowercase_ ) _lowerCAmelCase:Any = np.dot(x.T , h - y ) / y.size _lowerCAmelCase:Dict = theta - alpha * gradient # updating the weights _lowerCAmelCase:Dict = np.dot(lowercase_ , lowercase_ ) _lowerCAmelCase:Dict = sigmoid_function(lowercase_ ) _lowerCAmelCase:Any = cost_function(lowercase_ , lowercase_ ) if iterations % 100 == 0: print(F'loss: {j} \t' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": UpperCamelCase__ = datasets.load_iris() UpperCamelCase__ = iris.data[:, :2] UpperCamelCase__ = (iris.target != 0) * 1 UpperCamelCase__ = 0.1 UpperCamelCase__ = logistic_reg(alpha, x, y, max_iterations=7_0_0_0_0) print('''theta: ''', theta) # printing the theta i.e our weights vector def UpperCAmelCase ( snake_case : Optional[Any] ): return sigmoid_function( np.dot(lowercase_ , lowercase_ ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(1_0, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') (UpperCamelCase__) = (x[:, 0].min(), x[:, 0].max()) (UpperCamelCase__) = (x[:, 1].min(), x[:, 1].max()) (UpperCamelCase__) = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) UpperCamelCase__ = np.c_[xxa.ravel(), xxa.ravel()] UpperCamelCase__ = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
227
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class A__ : '''simple docstring''' snake_case__ = LEDConfig snake_case__ = {} snake_case__ = """gelu""" def __init__( self : List[str] , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Optional[int]=13 , _SCREAMING_SNAKE_CASE : Any=7 , _SCREAMING_SNAKE_CASE : str=True , _SCREAMING_SNAKE_CASE : int=False , _SCREAMING_SNAKE_CASE : Dict=99 , _SCREAMING_SNAKE_CASE : Dict=32 , _SCREAMING_SNAKE_CASE : Optional[Any]=2 , _SCREAMING_SNAKE_CASE : Tuple=4 , _SCREAMING_SNAKE_CASE : Dict=37 , _SCREAMING_SNAKE_CASE : Optional[int]=0.1 , _SCREAMING_SNAKE_CASE : List[Any]=0.1 , _SCREAMING_SNAKE_CASE : Tuple=20 , _SCREAMING_SNAKE_CASE : Dict=2 , _SCREAMING_SNAKE_CASE : Tuple=1 , _SCREAMING_SNAKE_CASE : int=0 , _SCREAMING_SNAKE_CASE : Any=4 , ): """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = seq_length UpperCamelCase = is_training UpperCamelCase = use_labels UpperCamelCase = vocab_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = eos_token_id UpperCamelCase = pad_token_id UpperCamelCase = bos_token_id UpperCamelCase = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after UpperCamelCase = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests UpperCamelCase = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def _SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" UpperCamelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCamelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCamelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) UpperCamelCase = prepare_led_inputs_dict(snake_case_ , snake_case_ , snake_case_ ) UpperCamelCase = tf.concat( [tf.zeros_like(snake_case_ )[:, :-1], tf.ones_like(snake_case_ )[:, -1:]] , axis=-1 , ) UpperCamelCase = global_attention_mask return config, inputs_dict def _SCREAMING_SNAKE_CASE ( self : List[Any] , _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : str ): """simple docstring""" UpperCamelCase = TFLEDModel(config=snake_case_ ).get_decoder() UpperCamelCase = inputs_dict['input_ids'] UpperCamelCase = input_ids[:1, :] UpperCamelCase = inputs_dict['attention_mask'][:1, :] UpperCamelCase = 1 # first forward pass UpperCamelCase = model(snake_case_ , attention_mask=snake_case_ , use_cache=snake_case_ ) UpperCamelCase , UpperCamelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCamelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCamelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCamelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCamelCase = model(snake_case_ , attention_mask=snake_case_ )[0] UpperCamelCase = model(snake_case_ , attention_mask=snake_case_ , past_key_values=snake_case_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCamelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCamelCase = output_from_no_past[:, -3:, random_slice_idx] UpperCamelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(snake_case_ , snake_case_ , rtol=1E-3 ) def lowercase__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=None , _UpperCamelCase=None , _UpperCamelCase=None , _UpperCamelCase=None , ) -> Optional[Any]: """simple docstring""" if attention_mask is None: UpperCamelCase = tf.cast(tf.math.not_equal(lowercase_ , config.pad_token_id) , tf.inta) if decoder_attention_mask is None: UpperCamelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id) , tf.inta), ] , axis=-1 , ) if head_mask is None: UpperCamelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads)) if decoder_head_mask is None: UpperCamelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads)) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class A__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case__ = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () snake_case__ = (TFLEDForConditionalGeneration,) if is_tf_available() else () snake_case__ = ( { """conversational""": TFLEDForConditionalGeneration, """feature-extraction""": TFLEDModel, """summarization""": TFLEDForConditionalGeneration, """text2text-generation""": TFLEDForConditionalGeneration, """translation""": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) snake_case__ = True snake_case__ = False snake_case__ = False snake_case__ = False def _SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" UpperCamelCase = TFLEDModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=snake_case_ ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*snake_case_ ) def _SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = tf.zeros_like(inputs_dict['attention_mask'] ) UpperCamelCase = 2 UpperCamelCase = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict['global_attention_mask'] , ) UpperCamelCase = True UpperCamelCase = self.model_tester.seq_length UpperCamelCase = self.model_tester.encoder_seq_length def check_decoder_attentions_output(_SCREAMING_SNAKE_CASE : Union[str, Any] ): UpperCamelCase = outputs.decoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(_SCREAMING_SNAKE_CASE : Optional[int] ): UpperCamelCase = [t.numpy() for t in outputs.encoder_attentions] UpperCamelCase = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: UpperCamelCase = True UpperCamelCase = False UpperCamelCase = False UpperCamelCase = model_class(snake_case_ ) UpperCamelCase = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) UpperCamelCase = len(snake_case_ ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) if self.is_encoder_decoder: UpperCamelCase = model_class(snake_case_ ) UpperCamelCase = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_decoder_attentions_output(snake_case_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCamelCase = True UpperCamelCase = model_class(snake_case_ ) UpperCamelCase = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) # Check attention is always last and order is fine UpperCamelCase = True UpperCamelCase = True UpperCamelCase = model_class(snake_case_ ) UpperCamelCase = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(snake_case_ ) ) self.assertEqual(model.config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) @unittest.skip('LED keeps using potentially symbolic tensors in conditionals and breaks tracing.' ) def _SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" pass def _SCREAMING_SNAKE_CASE ( self : List[Any] ): """simple docstring""" pass def lowercase__ ( _UpperCamelCase) -> Optional[int]: """simple docstring""" return tf.constant(lowercase_ , dtype=tf.intaa) __magic_name__ : List[Any] = 1e-4 @slow @require_tf class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" UpperCamelCase = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ).led # change to intended input here UpperCamelCase = _long_tensor([512 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) UpperCamelCase = _long_tensor([128 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) UpperCamelCase = prepare_led_inputs_dict(model.config , snake_case_ , snake_case_ ) UpperCamelCase = model(**snake_case_ )[0] UpperCamelCase = (1, 1024, 768) self.assertEqual(output.shape , snake_case_ ) # change to expected output here UpperCamelCase = tf.convert_to_tensor( [[2.3_0_5_0, 2.8_2_7_9, 0.6_5_3_1], [-1.8_4_5_7, -0.1_4_5_5, -3.5_6_6_1], [-1.0_1_8_6, 0.4_5_8_6, -2.2_0_4_3]] , ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-3 ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" UpperCamelCase = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ) # change to intended input here UpperCamelCase = _long_tensor([512 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) UpperCamelCase = _long_tensor([128 * [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69]] ) UpperCamelCase = prepare_led_inputs_dict(model.config , snake_case_ , snake_case_ ) UpperCamelCase = model(**snake_case_ )[0] UpperCamelCase = (1, 1024, model.config.vocab_size) self.assertEqual(output.shape , snake_case_ ) # change to expected output here UpperCamelCase = tf.convert_to_tensor( [[3_3.6_5_0_7, 6.4_5_7_2, 1_6.8_0_8_9], [5.8_7_3_9, -2.4_2_3_8, 1_1.2_9_0_2], [-3.2_1_3_9, -4.3_1_4_9, 4.2_7_8_3]] , ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-3 , rtol=1E-3 )
280
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
0
'''simple docstring''' import copy import re class _lowerCAmelCase : __SCREAMING_SNAKE_CASE : Union[str, Any] = 'hp' __SCREAMING_SNAKE_CASE : Tuple = {} __SCREAMING_SNAKE_CASE : Optional[int] = None @classmethod def _a (cls , lowercase , lowercase ): A_ : int = prefix A_ : Tuple = defaults cls.build_naming_info() @staticmethod def _a (lowercase , lowercase ): if len(snake_case_ ) == 0: return "" A_ : Optional[Any] = None if any(char.isdigit() for char in word ): raise Exception(F'Parameters should not contain numbers: \'{word}\' contains a number' ) if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(snake_case_ ) + 1 ): A_ : str = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: A_ : List[str] = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(lowercase ): A_ : Optional[Any] = """""" while integer != 0: A_ : Dict = chr(ord("""A""" ) + integer % 10 ) + s integer //= 10 return s A_ : List[str] = 0 while True: A_ : Dict = word + """#""" + int_to_alphabetic(snake_case_ ) if sword in info["reverse_short_word"]: continue else: A_ : int = sword break A_ : Union[str, Any] = short_word A_ : Union[str, Any] = word return short_word @staticmethod def _a (lowercase , lowercase ): A_ : Tuple = param_name.split("""_""" ) A_ : Optional[Any] = [TrialShortNamer.shortname_for_word(snake_case_ , snake_case_ ) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name A_ : Any = ["""""", """_"""] for separator in separators: A_ : str = separator.join(snake_case_ ) if shortname not in info["reverse_short_param"]: A_ : int = shortname A_ : Dict = param_name return shortname return param_name @staticmethod def _a (lowercase , lowercase ): A_ : Dict = TrialShortNamer.shortname_for_key(snake_case_ , snake_case_ ) A_ : Tuple = short_name A_ : Optional[int] = param_name @classmethod def _a (cls ): if cls.NAMING_INFO is not None: return A_ : List[Any] = { """short_word""": {}, """reverse_short_word""": {}, """short_param""": {}, """reverse_short_param""": {}, } A_ : int = list(cls.DEFAULTS.keys() ) for k in field_keys: cls.add_new_param_name(snake_case_ , snake_case_ ) A_ : List[str] = info @classmethod def _a (cls , lowercase ): cls.build_naming_info() assert cls.PREFIX is not None A_ : List[Any] = [copy.copy(cls.PREFIX )] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(F'You should provide a default value for the param name {k} with value {v}' ) if v == cls.DEFAULTS[k]: # The default value is not added to the name continue A_ : Dict = cls.NAMING_INFO["""short_param"""][k] if isinstance(snake_case_ , snake_case_ ): A_ : Tuple = 1 if v else 0 A_ : Any = """""" if isinstance(snake_case_ , (int, float) ) else """-""" A_ : Optional[Any] = F'{key}{sep}{v}' name.append(snake_case_ ) return "_".join(snake_case_ ) @classmethod def _a (cls , lowercase ): A_ : List[Any] = repr[len(cls.PREFIX ) + 1 :] if repr == "": A_ : Optional[int] = [] else: A_ : Any = repr.split("""_""" ) A_ : List[Any] = {} for value in values: if "-" in value: A_, A_ : Optional[int] = value.split("""-""" ) else: A_ : List[Any] = re.sub("""[0-9.]""" , """""" , snake_case_ ) A_ : Union[str, Any] = float(re.sub("""[^0-9.]""" , """""" , snake_case_ ) ) A_ : Union[str, Any] = cls.NAMING_INFO["""reverse_short_param"""][p_k] A_ : List[str] = p_v for k in cls.DEFAULTS: if k not in parameters: A_ : Any = cls.DEFAULTS[k] return parameters
667
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
0
'''simple docstring''' def a_ ( UpperCamelCase_ = 1_0_0_0 ): return sum(e for e in range(3 , lowercase_ ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F"{solution() = }")
452
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
0
"""simple docstring""" import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case_ : def __init__( self , a_ , a_=1_3 , a_=[3_0, 3_0] , a_=2 , a_=3 , a_=True , a_=True , a_=3_2 , a_=5 , a_=4 , a_=3_7 , a_="gelu" , a_=0.1 , a_=0.1 , a_=1_0 , a_=0.02 , a_=3 , a_=None , a_=8 , a_=1_0 , ): a_ : Tuple = parent a_ : int = batch_size a_ : int = image_size a_ : Optional[Any] = patch_size a_ : Optional[Any] = num_channels a_ : str = is_training a_ : str = use_labels a_ : List[str] = hidden_size a_ : List[str] = num_hidden_layers a_ : Tuple = num_attention_heads a_ : Optional[int] = intermediate_size a_ : Any = hidden_act a_ : Union[str, Any] = hidden_dropout_prob a_ : str = attention_probs_dropout_prob a_ : List[Any] = type_sequence_label_size a_ : Any = initializer_range a_ : int = num_labels a_ : int = scope a_ : List[str] = n_targets a_ : Optional[int] = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens a_ : List[Any] = (image_size[1] // patch_size) * (image_size[0] // patch_size) a_ : int = num_patches + 1 + self.num_detection_tokens def snake_case_ ( self ): a_ : int = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) a_ : str = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) a_ : int = [] for i in range(self.batch_size ): a_ : Union[str, Any] = {} a_ : Tuple = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=snake_case_ ) a_ : int = torch.rand(self.n_targets , 4 , device=snake_case_ ) labels.append(snake_case_ ) a_ : List[Any] = self.get_config() return config, pixel_values, labels def snake_case_ ( self ): return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def snake_case_ ( self , a_ , a_ , a_ ): a_ : List[Any] = YolosModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() a_ : Any = model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def snake_case_ ( self , a_ , a_ , a_ ): a_ : Dict = YolosForObjectDetection(snake_case_ ) model.to(snake_case_ ) model.eval() a_ : Dict = model(pixel_values=snake_case_ ) a_ : int = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) a_ : Dict = model(pixel_values=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def snake_case_ ( self ): a_ : Any = self.prepare_config_and_inputs() a_ , a_ , a_ : Any = config_and_inputs a_ : int = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class snake_case_ ( __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,unittest.TestCase ): __lowerCAmelCase = (YolosModel, YolosForObjectDetection) if is_torch_available() else () __lowerCAmelCase = ( {"feature-extraction": YolosModel, "object-detection": YolosForObjectDetection} if is_torch_available() else {} ) __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def snake_case_ ( self , a_ , a_ , a_=False ): a_ : int = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": a_ : str = [] for i in range(self.model_tester.batch_size ): a_ : int = {} a_ : List[Any] = torch.ones( size=(self.model_tester.n_targets,) , device=snake_case_ , dtype=torch.long ) a_ : int = torch.ones( self.model_tester.n_targets , 4 , device=snake_case_ , dtype=torch.float ) labels.append(snake_case_ ) a_ : Dict = labels return inputs_dict def snake_case_ ( self ): a_ : List[Any] = YolosModelTester(self ) a_ : Any = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=3_7 ) def snake_case_ ( self ): self.config_tester.run_common_tests() def snake_case_ ( self ): # YOLOS does not use inputs_embeds pass def snake_case_ ( self ): a_ , a_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a_ : int = model_class(snake_case_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) a_ : str = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case_ , nn.Linear ) ) def snake_case_ ( self ): a_ , a_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a_ : List[Any] = model_class(snake_case_ ) a_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a_ : List[str] = [*signature.parameters.keys()] a_ : str = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case_ ) def snake_case_ ( self ): a_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def snake_case_ ( self ): a_ , a_ : Any = self.model_tester.prepare_config_and_inputs_for_common() a_ : Tuple = True # in YOLOS, the seq_len is different a_ : List[Any] = self.model_tester.expected_seq_len for model_class in self.all_model_classes: a_ : Any = True a_ : int = False a_ : List[Any] = True a_ : Union[str, Any] = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): a_ : List[str] = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) a_ : List[str] = outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] a_ : List[str] = True a_ : int = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): a_ : Dict = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) a_ : int = outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) a_ : List[str] = len(snake_case_ ) # Check attention is always last and order is fine a_ : Optional[Any] = True a_ : Optional[Any] = True a_ : Dict = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): a_ : int = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) a_ : Optional[Any] = 1 self.assertEqual(out_len + added_hidden_states , len(snake_case_ ) ) a_ : Optional[int] = outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def snake_case_ ( self ): def check_hidden_states_output(a_ , a_ , a_ ): a_ : List[Any] = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): a_ : Dict = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) a_ : Tuple = outputs.hidden_states a_ : Any = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(snake_case_ ) , snake_case_ ) # YOLOS has a different seq_length a_ : Optional[int] = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) a_ , a_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a_ : List[str] = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a_ : Dict = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) def snake_case_ ( self ): a_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*snake_case_ ) @slow def snake_case_ ( self ): for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a_ : Optional[Any] = YolosModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def lowerCAmelCase_ ( ) -> Tuple: a_ : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class snake_case_ ( unittest.TestCase ): @cached_property def snake_case_ ( self ): return AutoImageProcessor.from_pretrained("hustvl/yolos-small" ) if is_vision_available() else None @slow def snake_case_ ( self ): a_ : int = YolosForObjectDetection.from_pretrained("hustvl/yolos-small" ).to(snake_case_ ) a_ : Optional[Any] = self.default_image_processor a_ : Optional[int] = prepare_img() a_ : Optional[int] = image_processor(images=snake_case_ , return_tensors="pt" ).to(snake_case_ ) # forward pass with torch.no_grad(): a_ : List[Any] = model(inputs.pixel_values ) # verify outputs a_ : Tuple = torch.Size((1, 1_0_0, 9_2) ) self.assertEqual(outputs.logits.shape , snake_case_ ) a_ : Optional[int] = torch.tensor( [[-24.0_248, -10.3_024, -14.8_290], [-42.0_392, -16.8_200, -27.4_334], [-27.2_743, -11.8_154, -18.7_148]] , device=snake_case_ , ) a_ : int = torch.tensor( [[0.2_559, 0.5_455, 0.4_706], [0.2_989, 0.7_279, 0.1_875], [0.7_732, 0.4_017, 0.4_462]] , device=snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , snake_case_ , atol=1e-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , snake_case_ , atol=1e-4 ) ) # verify postprocessing a_ : List[Any] = image_processor.post_process_object_detection( snake_case_ , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] a_ : Any = torch.tensor([0.9_994, 0.9_790, 0.9_964, 0.9_972, 0.9_861] ).to(snake_case_ ) a_ : Union[str, Any] = [7_5, 7_5, 1_7, 6_3, 1_7] a_ : Optional[int] = torch.tensor([3_3_5.0_6_0_9, 79.3_848, 3_7_5.4_2_1_6, 1_8_7.2_4_9_5] ).to(snake_case_ ) self.assertEqual(len(results["scores"] ) , 5 ) self.assertTrue(torch.allclose(results["scores"] , snake_case_ , atol=1e-4 ) ) self.assertSequenceEqual(results["labels"].tolist() , snake_case_ ) self.assertTrue(torch.allclose(results["boxes"][0, :] , snake_case_ ) )
237
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
0
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler UpperCamelCase = 16 UpperCamelCase = 32 def _lowerCamelCase ( UpperCAmelCase_ : Accelerator, UpperCAmelCase_ : int = 16, UpperCAmelCase_ : str = "bert-base-cased" ) -> Any: """simple docstring""" A__ = AutoTokenizer.from_pretrained(lowercase_ ) A__ = load_dataset("glue", "mrpc" ) def tokenize_function(UpperCAmelCase_ : Any ): # max_length=None => use the model max length (it's actually the default) A__ = tokenizer(examples["sentence1"], examples["sentence2"], truncation=lowercase_, max_length=lowercase_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset A__ = datasets.map( lowercase_, batched=lowercase_, remove_columns=["idx", "sentence1", "sentence2"], load_from_cache_file=lowercase_ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library A__ = tokenized_datasets.rename_column("label", "labels" ) def collate_fn(UpperCAmelCase_ : Union[str, Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase_, padding="max_length", max_length=128, return_tensors="pt" ) return tokenizer.pad(lowercase_, padding="longest", return_tensors="pt" ) # Instantiate dataloaders. A__ = DataLoader( tokenized_datasets["train"], shuffle=lowercase_, collate_fn=lowercase_, batch_size=lowercase_ ) A__ = DataLoader( tokenized_datasets["validation"], shuffle=lowercase_, collate_fn=lowercase_, batch_size=lowercase_ ) return train_dataloader, eval_dataloader def _lowerCamelCase ( UpperCAmelCase_ : Optional[int], UpperCAmelCase_ : Optional[int] ) -> Tuple: """simple docstring""" A__ = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs A__ = config["lr"] A__ = int(config["num_epochs"] ) A__ = int(config["seed"] ) A__ = int(config["batch_size"] ) A__ = args.model_name_or_path set_seed(lowercase_ ) A__ , A__ = get_dataloaders(lowercase_, lowercase_, lowercase_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) A__ = AutoModelForSequenceClassification.from_pretrained(lowercase_, return_dict=lowercase_ ) # Instantiate optimizer A__ = ( AdamW if accelerator.state.deepspeed_plugin is None or "optimizer" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) A__ = optimizer_cls(params=model.parameters(), lr=lowercase_ ) if accelerator.state.deepspeed_plugin is not None: A__ = accelerator.state.deepspeed_plugin.deepspeed_config[ "gradient_accumulation_steps" ] else: A__ = 1 A__ = (len(lowercase_ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): A__ = get_linear_schedule_with_warmup( optimizer=lowercase_, num_warmup_steps=0, num_training_steps=lowercase_, ) else: A__ = DummyScheduler(lowercase_, total_num_steps=lowercase_, warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. A__ , A__ , A__ , A__ , A__ = accelerator.prepare( lowercase_, lowercase_, lowercase_, lowercase_, lowercase_ ) # We need to keep track of how many total steps we have iterated over A__ = 0 # We also need to keep track of the stating epoch so files are named properly A__ = 0 # Now we train the model A__ = evaluate.load("glue", "mrpc" ) A__ = 0 A__ = {} for epoch in range(lowercase_, lowercase_ ): model.train() for step, batch in enumerate(lowercase_ ): A__ = model(**lowercase_ ) A__ = outputs.loss A__ = loss / gradient_accumulation_steps accelerator.backward(lowercase_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() A__ = 0 for step, batch in enumerate(lowercase_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): A__ = model(**lowercase_ ) A__ = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times A__ , A__ = accelerator.gather( (predictions, batch["labels"]) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase_ ) - 1: A__ = predictions[: len(eval_dataloader.dataset ) - samples_seen] A__ = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase_, references=lowercase_, ) A__ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""", lowercase_ ) A__ = eval_metric["accuracy"] if best_performance < eval_metric["accuracy"]: A__ = eval_metric["accuracy"] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), F"""Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}""" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir, "all_results.json" ), "w" ) as f: json.dump(lowercase_, lowercase_ ) def _lowerCamelCase ( ) -> List[Any]: """simple docstring""" A__ = argparse.ArgumentParser(description="Simple example of training script tracking peak GPU memory usage." ) parser.add_argument( "--model_name_or_path", type=lowercase_, default="bert-base-cased", help="Path to pretrained model or model identifier from huggingface.co/models.", required=lowercase_, ) parser.add_argument( "--output_dir", type=lowercase_, default=".", help="Optional save directory where all checkpoint folders will be stored. Default is the current working directory.", ) parser.add_argument( "--performance_lower_bound", type=lowercase_, default=lowercase_, help="Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.", ) parser.add_argument( "--num_epochs", type=lowercase_, default=3, help="Number of train epochs.", ) A__ = parser.parse_args() A__ = {"lr": 2e-5, "num_epochs": args.num_epochs, "seed": 42, "batch_size": 16} training_function(lowercase_, lowercase_ ) if __name__ == "__main__": main()
104
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class lowerCamelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = CTRLTokenizer SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _a (self ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCAmelCase__ : Any = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] UpperCAmelCase__ : int = dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) UpperCAmelCase__ : str = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] UpperCAmelCase__ : List[str] = {"""unk_token""": """<unk>"""} UpperCAmelCase__ : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(snake_case_ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(snake_case_ ) ) def _a (self , **_lowerCamelCase ): """simple docstring""" kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **snake_case_ ) def _a (self , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Dict = """adapt react readapt apt""" UpperCAmelCase__ : int = """adapt react readapt apt""" return input_text, output_text def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCAmelCase__ : List[Any] = """adapt react readapt apt""" UpperCAmelCase__ : str = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() UpperCAmelCase__ : Any = tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) UpperCAmelCase__ : List[str] = tokens + [tokenizer.unk_token] UpperCAmelCase__ : Optional[Any] = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case_ ) , snake_case_ )
182
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
0
import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : int = [ 'decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(lowercase_ , lowercase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase , _lowercase : Optional[Any] = emb.weight.shape _lowercase : Tuple = nn.Linear(lowercase_ , lowercase_ , bias=lowercase_ ) _lowercase : Dict = emb.weight.data return lin_layer def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Optional[int] = torch.load(lowercase_ , map_location='cpu' ) _lowercase : List[Any] = Namespace(**checkpoint['cfg']['model'] ) _lowercase : str = checkpoint['model'] remove_ignore_keys_(lowercase_ ) _lowercase : Optional[Any] = state_dict['decoder.embed_tokens.weight'].shape[0] _lowercase : Optional[Any] = {key.replace('decoder' , 'model' ): val for key, val in state_dict.items()} _lowercase : Optional[Any] = XGLMConfig( vocab_size=lowercase_ , max_position_embeddings=args.max_target_positions , num_layers=args.decoder_layers , attention_heads=args.decoder_attention_heads , ffn_dim=args.decoder_ffn_embed_dim , d_model=args.decoder_embed_dim , layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='gelu' , scale_embedding=not args.no_scale_embedding , tie_word_embeddings=args.share_decoder_input_output_embed , ) _lowercase : Optional[int] = XGLMForCausalLM(lowercase_ ) _lowercase : Union[str, Any] = model.load_state_dict(lowercase_ , strict=lowercase_ ) print(lowercase_ ) _lowercase : str = make_linear_from_emb(model.model.embed_tokens ) return model if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") SCREAMING_SNAKE_CASE : Dict = parser.parse_args() SCREAMING_SNAKE_CASE : Optional[int] = convert_fairseq_xglm_checkpoint_from_disk(args.fairseq_path) model.save_pretrained(args.pytorch_dump_folder_path)
89
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class __A : """simple docstring""" def __init__( self , a__ , a__=2 , a__=True , a__=False , a__=10 , a__=3 , a__=32 * 4 , a__=32 * 6 , a__=4 , a__=32 , ): """simple docstring""" _lowerCamelCase : Tuple = parent _lowerCamelCase : Any = batch_size _lowerCamelCase : Optional[int] = is_training _lowerCamelCase : Optional[int] = use_auxiliary_loss _lowerCamelCase : Any = num_queries _lowerCamelCase : Any = num_channels _lowerCamelCase : List[Any] = min_size _lowerCamelCase : Any = max_size _lowerCamelCase : str = num_labels _lowerCamelCase : Union[str, Any] = mask_feature_size def __snake_case ( self): """simple docstring""" _lowerCamelCase : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size]).to( snake_case_) _lowerCamelCase : int = torch.ones([self.batch_size, self.min_size, self.max_size] , device=snake_case_) _lowerCamelCase : str = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=snake_case_) > 0.5 ).float() _lowerCamelCase : int = (torch.rand((self.batch_size, self.num_labels) , device=snake_case_) > 0.5).long() _lowerCamelCase : str = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __snake_case ( self): """simple docstring""" return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def __snake_case ( self): """simple docstring""" _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Dict = self.prepare_config_and_inputs() _lowerCamelCase : Optional[int] = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def __snake_case ( self , a__ , a__): """simple docstring""" _lowerCamelCase : List[Any] = output.encoder_hidden_states _lowerCamelCase : Optional[Any] = output.pixel_decoder_hidden_states _lowerCamelCase : Dict = output.transformer_decoder_hidden_states self.parent.assertTrue(len(snake_case_) , len(config.backbone_config.depths)) self.parent.assertTrue(len(snake_case_) , len(config.backbone_config.depths)) self.parent.assertTrue(len(snake_case_) , config.decoder_config.decoder_layers) def __snake_case ( self , a__ , a__ , a__ , a__=False): """simple docstring""" with torch.no_grad(): _lowerCamelCase : Union[str, Any] = MaskFormerModel(config=snake_case_) model.to(snake_case_) model.eval() _lowerCamelCase : Optional[int] = model(pixel_values=snake_case_ , pixel_mask=snake_case_) _lowerCamelCase : int = model(snake_case_ , output_hidden_states=snake_case_) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(output.encoder_last_hidden_state is not None) if output_hidden_states: self.check_output_hidden_state(snake_case_ , snake_case_) def __snake_case ( self , a__ , a__ , a__ , a__ , a__): """simple docstring""" _lowerCamelCase : List[Any] = MaskFormerForInstanceSegmentation(config=snake_case_) model.to(snake_case_) model.eval() def comm_check_on_output(a__): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(result.encoder_last_hidden_state is not None) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1)) with torch.no_grad(): _lowerCamelCase : Tuple = model(pixel_values=snake_case_ , pixel_mask=snake_case_) _lowerCamelCase : Any = model(snake_case_) comm_check_on_output(snake_case_) _lowerCamelCase : int = model( pixel_values=snake_case_ , pixel_mask=snake_case_ , mask_labels=snake_case_ , class_labels=snake_case_) comm_check_on_output(snake_case_) self.parent.assertTrue(result.loss is not None) self.parent.assertEqual(result.loss.shape , torch.Size([1])) @require_torch class __A ( __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,unittest.TestCase ): """simple docstring""" UpperCAmelCase__ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () UpperCAmelCase__ = ( {"""feature-extraction""": MaskFormerModel, """image-segmentation""": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def __snake_case ( self): """simple docstring""" _lowerCamelCase : Optional[int] = MaskFormerModelTester(self) _lowerCamelCase : Optional[Any] = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_) def __snake_case ( self): """simple docstring""" self.config_tester.run_common_tests() def __snake_case ( self): """simple docstring""" _lowerCamelCase, _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(snake_case_ , **snake_case_ , output_hidden_states=snake_case_) def __snake_case ( self): """simple docstring""" _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*snake_case_) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''') def __snake_case ( self): """simple docstring""" pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''') def __snake_case ( self): """simple docstring""" pass @unittest.skip(reason='''MaskFormer is not a generative model''') def __snake_case ( self): """simple docstring""" pass @unittest.skip(reason='''MaskFormer does not use token embeddings''') def __snake_case ( self): """simple docstring""" pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''') def __snake_case ( self): """simple docstring""" pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''') def __snake_case ( self): """simple docstring""" pass def __snake_case ( self): """simple docstring""" _lowerCamelCase, _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : List[Any] = model_class(snake_case_) _lowerCamelCase : Dict = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase : List[Any] = [*signature.parameters.keys()] _lowerCamelCase : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , snake_case_) @slow def __snake_case ( self): """simple docstring""" for model_name in ["facebook/maskformer-swin-small-coco"]: _lowerCamelCase : Optional[int] = MaskFormerModel.from_pretrained(snake_case_) self.assertIsNotNone(snake_case_) def __snake_case ( self): """simple docstring""" _lowerCamelCase : List[str] = (self.model_tester.min_size,) * 2 _lowerCamelCase : Union[str, Any] = { '''pixel_values''': torch.randn((2, 3, *size) , device=snake_case_), '''mask_labels''': torch.randn((2, 10, *size) , device=snake_case_), '''class_labels''': torch.zeros(2 , 10 , device=snake_case_).long(), } _lowerCamelCase : Optional[int] = MaskFormerForInstanceSegmentation(MaskFormerConfig()).to(snake_case_) _lowerCamelCase : List[Any] = model(**snake_case_) self.assertTrue(outputs.loss is not None) def __snake_case ( self): """simple docstring""" _lowerCamelCase, _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(snake_case_ , **snake_case_ , output_hidden_states=snake_case_) def __snake_case ( self): """simple docstring""" _lowerCamelCase, _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : str = model_class(snake_case_).to(snake_case_) _lowerCamelCase : List[str] = model(**snake_case_ , output_attentions=snake_case_) self.assertTrue(outputs.attentions is not None) def __snake_case ( self): """simple docstring""" if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _lowerCamelCase : Optional[Any] = self.all_model_classes[1] _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() _lowerCamelCase : Optional[Any] = model_class(snake_case_) model.to(snake_case_) model.train() _lowerCamelCase : Dict = model(snake_case_ , mask_labels=snake_case_ , class_labels=snake_case_).loss loss.backward() def __snake_case ( self): """simple docstring""" _lowerCamelCase : Dict = self.all_model_classes[1] _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() _lowerCamelCase : Union[str, Any] = True _lowerCamelCase : Tuple = True _lowerCamelCase : Optional[Any] = model_class(snake_case_) model.to(snake_case_) model.train() _lowerCamelCase : Optional[int] = model(snake_case_ , mask_labels=snake_case_ , class_labels=snake_case_) _lowerCamelCase : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _lowerCamelCase : Any = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _lowerCamelCase : List[Any] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _lowerCamelCase : str = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=snake_case_) self.assertIsNotNone(encoder_hidden_states.grad) self.assertIsNotNone(pixel_decoder_hidden_states.grad) self.assertIsNotNone(transformer_decoder_hidden_states.grad) self.assertIsNotNone(attentions.grad) _lowerCamelCase = 1E-4 def __UpperCAmelCase( ): _lowerCamelCase : Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class __A ( unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self): """simple docstring""" return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''') if is_vision_available() else None ) def __snake_case ( self): """simple docstring""" _lowerCamelCase : Any = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''').to(snake_case_) _lowerCamelCase : Any = self.default_image_processor _lowerCamelCase : Tuple = prepare_img() _lowerCamelCase : Optional[Any] = image_processor(snake_case_ , return_tensors='''pt''').to(snake_case_) _lowerCamelCase : Tuple = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(snake_case_ , (1, 3, 800, 1088)) with torch.no_grad(): _lowerCamelCase : int = model(**snake_case_) _lowerCamelCase : int = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]]).to(snake_case_) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , snake_case_ , atol=snake_case_)) _lowerCamelCase : Optional[int] = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]]).to(snake_case_) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , snake_case_ , atol=snake_case_)) _lowerCamelCase : Optional[Any] = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]]).to(snake_case_) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , snake_case_ , atol=snake_case_)) def __snake_case ( self): """simple docstring""" _lowerCamelCase : int = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''') .to(snake_case_) .eval() ) _lowerCamelCase : Tuple = self.default_image_processor _lowerCamelCase : Union[str, Any] = prepare_img() _lowerCamelCase : List[Any] = image_processor(snake_case_ , return_tensors='''pt''').to(snake_case_) _lowerCamelCase : int = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(snake_case_ , (1, 3, 800, 1088)) with torch.no_grad(): _lowerCamelCase : Optional[int] = model(**snake_case_) # masks_queries_logits _lowerCamelCase : str = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _lowerCamelCase : Union[str, Any] = [ [-1.373_7124, -1.772_4937, -1.936_4233], [-1.597_7281, -1.986_7939, -2.152_3695], [-1.579_5398, -1.926_9832, -2.09_3942], ] _lowerCamelCase : List[str] = torch.tensor(snake_case_).to(snake_case_) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , snake_case_ , atol=snake_case_)) # class_queries_logits _lowerCamelCase : Optional[int] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1)) _lowerCamelCase : Optional[int] = torch.tensor( [ [1.6512e00, -5.2572e00, -3.3519e00], [3.6169e-02, -5.9025e00, -2.9313e00], [1.0766e-04, -7.7630e00, -5.1263e00], ]).to(snake_case_) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , snake_case_ , atol=snake_case_)) def __snake_case ( self): """simple docstring""" _lowerCamelCase : List[str] = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''') .to(snake_case_) .eval() ) _lowerCamelCase : Dict = self.default_image_processor _lowerCamelCase : Optional[Any] = prepare_img() _lowerCamelCase : Optional[Any] = image_processor(snake_case_ , return_tensors='''pt''').to(snake_case_) _lowerCamelCase : List[Any] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(snake_case_ , (1, 3, 800, 1088)) with torch.no_grad(): _lowerCamelCase : List[Any] = model(**snake_case_) # masks_queries_logits _lowerCamelCase : Optional[Any] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _lowerCamelCase : Dict = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] _lowerCamelCase : Optional[Any] = torch.tensor(snake_case_).to(snake_case_) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , snake_case_ , atol=snake_case_)) # class_queries_logits _lowerCamelCase : Tuple = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1)) _lowerCamelCase : List[Any] = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]]).to(snake_case_) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , snake_case_ , atol=snake_case_)) def __snake_case ( self): """simple docstring""" _lowerCamelCase : List[Any] = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''') .to(snake_case_) .eval() ) _lowerCamelCase : int = self.default_image_processor _lowerCamelCase : Optional[Any] = image_processor( [np.zeros((3, 800, 1333)), np.zeros((3, 800, 1333))] , segmentation_maps=[np.zeros((384, 384)).astype(np.floataa), np.zeros((384, 384)).astype(np.floataa)] , return_tensors='''pt''' , ) _lowerCamelCase : str = inputs['''pixel_values'''].to(snake_case_) _lowerCamelCase : Dict = [el.to(snake_case_) for el in inputs['''mask_labels''']] _lowerCamelCase : Any = [el.to(snake_case_) for el in inputs['''class_labels''']] with torch.no_grad(): _lowerCamelCase : List[Any] = model(**snake_case_) self.assertTrue(outputs.loss is not None)
114
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
0
import os from pathlib import Path def UpperCamelCase ( __lowercase : Tuple ,__lowercase : Union[str, Any] ,__lowercase : Optional[int] ,__lowercase : Dict ): '''simple docstring''' A_ : Union[str, Any] = { 'en': 'Machine learning is great, isn\'t it?', 'ru': 'Машинное обучение - это здорово, не так ли?', 'de': 'Maschinelles Lernen ist großartig, nicht wahr?', } # BLUE scores as follows: # "pair": [fairseq, transformers] A_ : str = { 'wmt16-en-de-dist-12-1': [28.3, 27.52], 'wmt16-en-de-dist-6-1': [27.4, 27.11], 'wmt16-en-de-12-1': [26.9, 25.75], } A_ : Dict = f'''{src_lang}-{tgt_lang}''' A_ : List[Any] = f'''\n---\nlanguage:\n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt16\n- allenai\nlicense: apache-2.0\ndatasets:\n- wmt16\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}.\n\nFor more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369).\n\nAll 3 models are available:\n\n* [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1)\n* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)\n* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)\n\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "allenai/{model_name}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n\n## Training data\n\nPretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369).\n\n## Eval results\n\nHere are the BLEU scores:\n\nmodel | fairseq | transformers\n-------|---------|----------\n{model_name} | {scores[model_name][0]} | {scores[model_name][1]}\n\nThe score is slightly below the score reported in the paper, as the researchers don\'t use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs.\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=5\nmkdir -p $DATA_DIR\nsacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt16/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)\n\n\n### BibTeX entry and citation info\n\n```\n@misc{{kasai2020deep,\n title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},\n author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},\n year={{2020}},\n eprint={{2006.10369}},\n archivePrefix={{arXiv}},\n primaryClass={{cs.CL}}\n}}\n```\n\n''' model_card_dir.mkdir(parents=lowercase_ ,exist_ok=lowercase_ ) A_ : List[Any] = os.path.join(lowercase_ ,'README.md' ) print(f'''Generating {path}''' ) with open(lowercase_ ,'w' ,encoding='utf-8' ) as f: f.write(lowercase_ ) # make sure we are under the root of the project _UpperCAmelCase = Path(__file__).resolve().parent.parent.parent _UpperCAmelCase = repo_dir / '''model_cards''' for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: _UpperCAmelCase = model_cards_dir / '''allenai''' / model_name write_model_card(model_card_dir, src_lang="""en""", tgt_lang="""de""", model_name=model_name)
558
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
0
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { '''EleutherAI/gpt-neo-1.3B''': '''https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json''', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class a_ (__SCREAMING_SNAKE_CASE ): __lowerCAmelCase : List[Any] = """gpt_neo""" __lowerCAmelCase : Dict = ["""past_key_values"""] __lowerCAmelCase : Tuple = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , snake_case_=5_0_2_5_7 , snake_case_=2_0_4_8 , snake_case_=2_0_4_8 , snake_case_=2_4 , snake_case_=[[["global", "local"], 1_2]] , snake_case_=1_6 , snake_case_=None , snake_case_=2_5_6 , snake_case_="gelu_new" , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.1 , snake_case_=1E-5 , snake_case_=0.02 , snake_case_=True , snake_case_=5_0_2_5_6 , snake_case_=5_0_2_5_6 , **snake_case_ , ): _lowerCAmelCase : Optional[int] = vocab_size _lowerCAmelCase : Dict = max_position_embeddings _lowerCAmelCase : Dict = hidden_size _lowerCAmelCase : int = num_layers _lowerCAmelCase : int = num_heads _lowerCAmelCase : Tuple = intermediate_size _lowerCAmelCase : Dict = window_size _lowerCAmelCase : List[str] = activation_function _lowerCAmelCase : str = resid_dropout _lowerCAmelCase : Tuple = embed_dropout _lowerCAmelCase : Union[str, Any] = attention_dropout _lowerCAmelCase : Optional[int] = classifier_dropout _lowerCAmelCase : Tuple = layer_norm_epsilon _lowerCAmelCase : List[str] = initializer_range _lowerCAmelCase : List[Any] = use_cache _lowerCAmelCase : Dict = bos_token_id _lowerCAmelCase : List[Any] = eos_token_id _lowerCAmelCase : List[Any] = attention_types _lowerCAmelCase : List[str] = self.expand_attention_types_params(snake_case_ ) if len(self.attention_layers ) != self.num_layers: raise ValueError( """Configuration for convolutional module is incorrect. """ """It is required that `len(config.attention_layers)` == `config.num_layers` """ f'but is `len(config.attention_layers) = {len(self.attention_layers )}`, ' f'`config.num_layers = {self.num_layers}`. ' """`config.attention_layers` is prepared using `config.attention_types`. """ """Please verify the value of `config.attention_types` argument.""" ) super().__init__(bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) @staticmethod def __UpperCamelCase ( snake_case_ ): _lowerCAmelCase : List[str] = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _UpperCAmelCase ( _lowerCamelCase : List[str] , _lowerCamelCase : str , _lowerCamelCase : str , _lowerCamelCase : Optional[int] ) -> str: import torch _lowerCAmelCase : Tuple = input.size() _lowerCAmelCase : int = len(lowercase_ ) _lowerCAmelCase : Union[str, Any] = shape[dimension] _lowerCAmelCase : Union[str, Any] = torch.arange(0 , lowercase_ , lowercase_ ) _lowerCAmelCase : Union[str, Any] = torch.div(sizedim - size , lowercase_ , rounding_mode="""floor""" ) + 1 _lowerCAmelCase : List[str] = torch.arange(lowercase_ ) + low_indices[:min_length][:, None] _lowerCAmelCase : Optional[Any] = [slice(lowercase_ )] * rank _lowerCAmelCase : Union[str, Any] = indices _lowerCAmelCase : List[str] = input[s] _lowerCAmelCase : List[Any] = list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowercase_ ) def _UpperCAmelCase ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any] ) -> List[str]: import torch _lowerCAmelCase : Tuple = torch.arange(1 , lowercase_ ) _lowerCAmelCase : Union[str, Any] = torch.remainder(lowercase_ , lowercase_ ) _lowerCAmelCase : Dict = remainders == 0 _lowerCAmelCase : Optional[Any] = candidates[divisor_indices] _lowerCAmelCase : Union[str, Any] = torch.max(lowercase_ ) return largest_divisor, torch.div(lowercase_ , lowercase_ , rounding_mode="""floor""" ) class a_ (__SCREAMING_SNAKE_CASE ): @property def __UpperCamelCase ( self ): _lowerCAmelCase : int = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction="""inputs""" ) _lowerCAmelCase : Optional[Any] = {0: """batch""", 1: """past_sequence + sequence"""} else: _lowerCAmelCase : str = {0: """batch""", 1: """sequence"""} return common_inputs @property def __UpperCamelCase ( self ): return self._config.num_heads def __UpperCamelCase ( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): _lowerCAmelCase : str = super(snake_case_ , self ).generate_dummy_inputs( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) # We need to order the input in the way they appears in the forward() _lowerCAmelCase : Dict = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch _lowerCAmelCase , _lowerCAmelCase : Dict = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values _lowerCAmelCase : Optional[int] = seqlen + 2 _lowerCAmelCase : List[Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCAmelCase : Union[str, Any] = [ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(self.num_layers ) ] _lowerCAmelCase : int = common_inputs["""attention_mask"""] if self.use_past: _lowerCAmelCase : List[Any] = ordered_inputs["""attention_mask"""].dtype _lowerCAmelCase : Tuple = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) return ordered_inputs @property def __UpperCamelCase ( self ): return 1_3
384
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
0
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class a__ : def __init__( self : Optional[Any] ,a__ : Any) -> int: """simple docstring""" _lowerCAmelCase:int = data _lowerCAmelCase:str = [0x6745_2301, 0xefcd_ab89, 0x98ba_dcfe, 0x1032_5476, 0xc3d2_e1f0] @staticmethod def __UpperCamelCase ( a__ : Optional[Any] ,a__ : Union[str, Any]) -> Optional[int]: """simple docstring""" return ((n << b) | (n >> (32 - b))) & 0xffff_ffff def __UpperCamelCase ( self : Any) -> int: """simple docstring""" _lowerCAmelCase:Tuple = B'''\x80''' + B'''\x00''' * (63 - (len(self.data) + 8) % 64) _lowerCAmelCase:str = self.data + padding + struct.pack('''>Q''' ,8 * len(self.data)) return padded_data def __UpperCamelCase ( self : Optional[int]) -> Optional[int]: """simple docstring""" return [ self.padded_data[i : i + 64] for i in range(0 ,len(self.padded_data) ,64) ] def __UpperCamelCase ( self : Optional[int] ,a__ : Tuple) -> List[Any]: """simple docstring""" _lowerCAmelCase:Dict = list(struct.unpack('''>16L''' ,snake_case_)) + [0] * 64 for i in range(16 ,80): _lowerCAmelCase:Any = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) ,1) return w def __UpperCamelCase ( self : int) -> Tuple: """simple docstring""" _lowerCAmelCase:Optional[int] = self.padding() _lowerCAmelCase:Dict = self.split_blocks() for block in self.blocks: _lowerCAmelCase:Union[str, Any] = self.expand_block(snake_case_) _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase:Optional[Any] = self.h for i in range(0 ,80): if 0 <= i < 20: _lowerCAmelCase:Dict = (b & c) | ((~b) & d) _lowerCAmelCase:Tuple = 0x5a82_7999 elif 20 <= i < 40: _lowerCAmelCase:Optional[int] = b ^ c ^ d _lowerCAmelCase:List[Any] = 0x6ed9_eba1 elif 40 <= i < 60: _lowerCAmelCase:List[str] = (b & c) | (b & d) | (c & d) _lowerCAmelCase:List[str] = 0x8f1b_bcdc elif 60 <= i < 80: _lowerCAmelCase:Optional[int] = b ^ c ^ d _lowerCAmelCase:int = 0xca62_c1d6 _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase:Optional[Any] = ( self.rotate(snake_case_ ,5) + f + e + k + expanded_block[i] & 0xffff_ffff, a, self.rotate(snake_case_ ,30), c, d, ) _lowerCAmelCase:Dict = ( self.h[0] + a & 0xffff_ffff, self.h[1] + b & 0xffff_ffff, self.h[2] + c & 0xffff_ffff, self.h[3] + d & 0xffff_ffff, self.h[4] + e & 0xffff_ffff, ) return ("{:08x}" * 5).format(*self.h) def UpperCAmelCase ( ): _lowerCAmelCase:Optional[int] = B'''Test String''' assert SHAaHash(lowercase_ ).final_hash() == hashlib.shaa(lowercase_ ).hexdigest() # noqa: S324 def UpperCAmelCase ( ): _lowerCAmelCase:Tuple = argparse.ArgumentParser(description='''Process some strings or files''' ) parser.add_argument( '''--string''' , dest='''input_string''' , default='''Hello World!! Welcome to Cryptography''' , help='''Hash the string''' , ) parser.add_argument('''--file''' , dest='''input_file''' , help='''Hash contents of a file''' ) _lowerCAmelCase:str = parser.parse_args() _lowerCAmelCase:str = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , '''rb''' ) as f: _lowerCAmelCase:Dict = f.read() else: _lowerCAmelCase:Union[str, Any] = bytes(lowercase_ , '''utf-8''' ) print(SHAaHash(lowercase_ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
227
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging __magic_name__ : Dict = logging.get_logger(__name__) __magic_name__ : Union[str, Any] = { '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/config.json''', # See all BART models at https://huggingface.co/models?filter=bart } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case__ = """bart""" snake_case__ = ["""past_key_values"""] snake_case__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : Dict , _SCREAMING_SNAKE_CASE : Union[str, Any]=5_0265 , _SCREAMING_SNAKE_CASE : List[Any]=1024 , _SCREAMING_SNAKE_CASE : List[str]=12 , _SCREAMING_SNAKE_CASE : List[Any]=4096 , _SCREAMING_SNAKE_CASE : Union[str, Any]=16 , _SCREAMING_SNAKE_CASE : int=12 , _SCREAMING_SNAKE_CASE : Tuple=4096 , _SCREAMING_SNAKE_CASE : Optional[int]=16 , _SCREAMING_SNAKE_CASE : List[str]=0.0 , _SCREAMING_SNAKE_CASE : List[Any]=0.0 , _SCREAMING_SNAKE_CASE : List[str]="gelu" , _SCREAMING_SNAKE_CASE : Union[str, Any]=1024 , _SCREAMING_SNAKE_CASE : Optional[int]=0.1 , _SCREAMING_SNAKE_CASE : str=0.0 , _SCREAMING_SNAKE_CASE : List[Any]=0.0 , _SCREAMING_SNAKE_CASE : Union[str, Any]=0.0_2 , _SCREAMING_SNAKE_CASE : str=0.0 , _SCREAMING_SNAKE_CASE : Optional[int]=False , _SCREAMING_SNAKE_CASE : Optional[int]=True , _SCREAMING_SNAKE_CASE : Tuple=3 , _SCREAMING_SNAKE_CASE : int=1 , _SCREAMING_SNAKE_CASE : Optional[int]=0 , _SCREAMING_SNAKE_CASE : str=2 , _SCREAMING_SNAKE_CASE : Optional[int]=True , _SCREAMING_SNAKE_CASE : Optional[int]=2 , _SCREAMING_SNAKE_CASE : str=2 , **_SCREAMING_SNAKE_CASE : Union[str, Any] , ): """simple docstring""" UpperCamelCase = vocab_size UpperCamelCase = max_position_embeddings UpperCamelCase = d_model UpperCamelCase = encoder_ffn_dim UpperCamelCase = encoder_layers UpperCamelCase = encoder_attention_heads UpperCamelCase = decoder_ffn_dim UpperCamelCase = decoder_layers UpperCamelCase = decoder_attention_heads UpperCamelCase = dropout UpperCamelCase = attention_dropout UpperCamelCase = activation_dropout UpperCamelCase = activation_function UpperCamelCase = init_std UpperCamelCase = encoder_layerdrop UpperCamelCase = decoder_layerdrop UpperCamelCase = classifier_dropout UpperCamelCase = use_cache UpperCamelCase = encoder_layers UpperCamelCase = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=snake_case_ , pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('force_bos_token_to_be_generated' , snake_case_ ): UpperCamelCase = self.bos_token_id warnings.warn( f'Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ' 'The config can simply be saved and uploaded again to be fixed.' ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: UpperCamelCase = {0: 'batch'} UpperCamelCase = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: UpperCamelCase = {0: 'batch', 1: 'decoder_sequence'} UpperCamelCase = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. UpperCamelCase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: UpperCamelCase , UpperCamelCase = self.num_layers for i in range(snake_case_ ): UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} else: UpperCamelCase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property def _SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = super().outputs else: UpperCamelCase = super(snake_case_ , self ).outputs if self.use_past: UpperCamelCase , UpperCamelCase = self.num_layers for i in range(snake_case_ ): UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def _SCREAMING_SNAKE_CASE ( self : int , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : List[Any] = -1 , _SCREAMING_SNAKE_CASE : Tuple = -1 , _SCREAMING_SNAKE_CASE : str = False , _SCREAMING_SNAKE_CASE : Tuple = None , ): """simple docstring""" UpperCamelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs UpperCamelCase = seq_length if not self.use_past else 1 UpperCamelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) UpperCamelCase = {f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} UpperCamelCase = dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch UpperCamelCase , UpperCamelCase = common_inputs['input_ids'].shape UpperCamelCase = common_inputs['decoder_input_ids'].shape[1] UpperCamelCase , UpperCamelCase = self.num_attention_heads UpperCamelCase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCamelCase = decoder_seq_length + 3 UpperCamelCase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) UpperCamelCase = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) UpperCamelCase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered UpperCamelCase , UpperCamelCase = self.num_layers UpperCamelCase = min(snake_case_ , snake_case_ ) UpperCamelCase = max(snake_case_ , snake_case_ ) - min_num_layers UpperCamelCase = 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. UpperCamelCase = encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _SCREAMING_SNAKE_CASE ( self : Tuple , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] = -1 , _SCREAMING_SNAKE_CASE : int = -1 , _SCREAMING_SNAKE_CASE : str = False , _SCREAMING_SNAKE_CASE : List[str] = None , ): """simple docstring""" UpperCamelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch UpperCamelCase , UpperCamelCase = common_inputs['input_ids'].shape # Not using the same length for past_key_values UpperCamelCase = seqlen + 2 UpperCamelCase , UpperCamelCase = self.num_layers UpperCamelCase , UpperCamelCase = self.num_attention_heads UpperCamelCase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCamelCase = common_inputs['attention_mask'].dtype UpperCamelCase = torch.cat( [common_inputs['attention_mask'], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) UpperCamelCase = [ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _SCREAMING_SNAKE_CASE ( self : List[Any] , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : str = -1 , _SCREAMING_SNAKE_CASE : Optional[Any] = -1 , _SCREAMING_SNAKE_CASE : Optional[int] = False , _SCREAMING_SNAKE_CASE : Any = None , ): """simple docstring""" UpperCamelCase = compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCamelCase = tokenizer.num_special_tokens_to_add(snake_case_ ) UpperCamelCase = compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence UpperCamelCase = [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size UpperCamelCase = dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _SCREAMING_SNAKE_CASE ( self : List[Any] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : int = -1 , _SCREAMING_SNAKE_CASE : List[Any] = -1 , _SCREAMING_SNAKE_CASE : Optional[Any] = False , _SCREAMING_SNAKE_CASE : Any = None , ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) elif self.task == "causal-lm": UpperCamelCase = self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: UpperCamelCase = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: UpperCamelCase = super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ )
280
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
0
'''simple docstring''' import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor if is_flax_available(): import jax from transformers import FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): def __init__(self , lowercase , lowercase=100 , lowercase=13 , lowercase=30 , lowercase=2 , lowercase=3 , lowercase=True , lowercase=True , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=10 , lowercase=0.02 , lowercase=3 , ): A_ : List[str] = parent A_ : Any = vocab_size A_ : Tuple = batch_size A_ : int = image_size A_ : Tuple = patch_size A_ : Tuple = num_channels A_ : Union[str, Any] = is_training A_ : int = use_labels A_ : Any = hidden_size A_ : Optional[Any] = num_hidden_layers A_ : List[str] = num_attention_heads A_ : Any = intermediate_size A_ : str = hidden_act A_ : List[Any] = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Optional[int] = type_sequence_label_size A_ : Optional[int] = initializer_range # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) A_ : Optional[Any] = (image_size // patch_size) ** 2 A_ : List[Any] = num_patches + 1 def _a (self ): A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : str = None if self.use_labels: A_ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Tuple = BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , ) return config, pixel_values, labels def _a (self , lowercase , lowercase , lowercase ): A_ : str = FlaxBeitModel(config=snake_case_ ) A_ : Tuple = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a (self , lowercase , lowercase , lowercase ): A_ : int = FlaxBeitForMaskedImageModeling(config=snake_case_ ) A_ : Dict = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def _a (self , lowercase , lowercase , lowercase ): A_ : Optional[Any] = self.type_sequence_label_size A_ : str = FlaxBeitForImageClassification(config=snake_case_ ) A_ : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A_ : str = 1 A_ : List[Any] = FlaxBeitForImageClassification(snake_case_ ) A_ : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A_ : Any = model(snake_case_ ) def _a (self ): A_ : Union[str, Any] = self.prepare_config_and_inputs() ( ( A_ ), ( A_ ), ( A_ ), ) : str = config_and_inputs A_ : Optional[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_flax class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): __SCREAMING_SNAKE_CASE : int = ( (FlaxBeitModel, FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling) if is_flax_available() else () ) def _a (self ): A_ : int = FlaxBeitModelTester(self ) A_ : Union[str, Any] = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def _a (self ): self.config_tester.run_common_tests() def _a (self ): A_, A_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : int = model_class(snake_case_ ) A_ : Dict = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Tuple = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , snake_case_ ) def _a (self ): A_, A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): A_ : List[str] = self._prepare_for_class(snake_case_ , snake_case_ ) A_ : Optional[Any] = model_class(snake_case_ ) @jax.jit def model_jitted(lowercase , **lowercase ): return model(pixel_values=snake_case_ , **snake_case_ ) with self.subTest("""JIT Enabled""" ): A_ : str = model_jitted(**snake_case_ ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): A_ : Union[str, Any] = model_jitted(**snake_case_ ).to_tuple() self.assertEqual(len(snake_case_ ) , len(snake_case_ ) ) for jitted_output, output in zip(snake_case_ , snake_case_ ): self.assertEqual(jitted_output.shape , output.shape ) def _a (self ): A_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def _a (self ): A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case_ ) def _a (self ): A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) @slow def _a (self ): for model_class_name in self.all_model_classes: A_ : List[str] = model_class_name.from_pretrained("""microsoft/beit-base-patch16-224""" ) A_ : Tuple = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(snake_case_ ) def a ( ): '''simple docstring''' A_ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @require_flax class _lowerCAmelCase ( unittest.TestCase ): @cached_property def _a (self ): return BeitImageProcessor.from_pretrained("""microsoft/beit-base-patch16-224""" ) if is_vision_available() else None @slow def _a (self ): A_ : Dict = FlaxBeitForMaskedImageModeling.from_pretrained("""microsoft/beit-base-patch16-224-pt22k""" ) A_ : Optional[Any] = self.default_image_processor A_ : str = prepare_img() A_ : List[Any] = image_processor(images=snake_case_ , return_tensors="""np""" ).pixel_values # prepare bool_masked_pos A_ : Tuple = np.ones((1, 196) , dtype=snake_case_ ) # forward pass A_ : str = model(pixel_values=snake_case_ , bool_masked_pos=snake_case_ ) A_ : Any = outputs.logits # verify the logits A_ : List[Any] = (1, 196, 8192) self.assertEqual(logits.shape , snake_case_ ) A_ : int = np.array( [[-3.24_37, 0.50_72, -13.91_74], [-3.24_56, 0.49_48, -13.94_01], [-3.20_33, 0.51_21, -13.85_50]] ) self.assertTrue(np.allclose(logits[bool_masked_pos][:3, :3] , snake_case_ , atol=1E-2 ) ) @slow def _a (self ): A_ : Optional[int] = FlaxBeitForImageClassification.from_pretrained("""microsoft/beit-base-patch16-224""" ) A_ : Optional[Any] = self.default_image_processor A_ : str = prepare_img() A_ : int = image_processor(images=snake_case_ , return_tensors="""np""" ) # forward pass A_ : List[Any] = model(**snake_case_ ) A_ : Optional[int] = outputs.logits # verify the logits A_ : str = (1, 1000) self.assertEqual(logits.shape , snake_case_ ) A_ : Any = np.array([-1.23_85, -1.09_87, -1.01_08] ) self.assertTrue(np.allclose(logits[0, :3] , snake_case_ , atol=1E-4 ) ) A_ : Optional[int] = 281 self.assertEqual(logits.argmax(-1 ).item() , snake_case_ ) @slow def _a (self ): A_ : Tuple = FlaxBeitForImageClassification.from_pretrained("""microsoft/beit-large-patch16-224-pt22k-ft22k""" ) A_ : Union[str, Any] = self.default_image_processor A_ : Dict = prepare_img() A_ : Union[str, Any] = image_processor(images=snake_case_ , return_tensors="""np""" ) # forward pass A_ : Dict = model(**snake_case_ ) A_ : List[str] = outputs.logits # verify the logits A_ : Dict = (1, 21841) self.assertEqual(logits.shape , snake_case_ ) A_ : List[str] = np.array([1.68_81, -0.27_87, 0.59_01] ) self.assertTrue(np.allclose(logits[0, :3] , snake_case_ , atol=1E-4 ) ) A_ : Optional[Any] = 2396 self.assertEqual(logits.argmax(-1 ).item() , snake_case_ )
667
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : Any = { '''configuration_clap''': [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapAudioConfig''', '''ClapConfig''', '''ClapTextConfig''', ], '''processing_clap''': ['''ClapProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Optional[int] = [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapModel''', '''ClapPreTrainedModel''', '''ClapTextModel''', '''ClapTextModelWithProjection''', '''ClapAudioModel''', '''ClapAudioModelWithProjection''', ] __SCREAMING_SNAKE_CASE : Optional[Any] = ['''ClapFeatureExtractor'''] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys __SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
452
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
0
"""simple docstring""" import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__=None, SCREAMING_SNAKE_CASE__=None ) -> Dict: return field(default_factory=lambda: default, metadata=lowercase_ ) @dataclass class snake_case_ : __lowerCAmelCase = list_field( default=[] ,metadata={ "help": ( "Model checkpoints to be provided to the AutoModel classes. Leave blank to benchmark the base version" " of all available models" ) } ,) __lowerCAmelCase = list_field( default=[8] ,metadata={"help": "List of batch sizes for which memory and time performance will be evaluated"} ) __lowerCAmelCase = list_field( default=[8, 3_2, 1_2_8, 5_1_2] ,metadata={"help": "List of sequence lengths for which memory and time performance will be evaluated"} ,) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Whether to benchmark inference of model. Inference can be disabled via --no-inference."} ,) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Whether to run on available cuda devices. Cuda can be disabled via --no-cuda."} ,) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Whether to run on available tpu devices. TPU can be disabled via --no-tpu."} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Use FP16 to accelerate inference."} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Benchmark training of model"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Verbose memory tracing"} ) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Whether to perform speed measurements. Speed measurements can be disabled via --no-speed."} ,) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={ "help": "Whether to perform memory measurements. Memory measurements can be disabled via --no-memory" } ,) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Trace memory line by line"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Save result to a CSV file"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Save all print statements in a log file"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Whether to print environment information"} ) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={ "help": ( "Whether to use multiprocessing for memory and speed measurement. It is highly recommended to use" " multiprocessing for accurate CPU and GPU memory measurements. This option should only be disabled" " for debugging / testing and on TPU." ) } ,) __lowerCAmelCase = field( default=f"""inference_time_{round(time() )}.csv""" ,metadata={"help": "CSV filename used if saving time results to csv."} ,) __lowerCAmelCase = field( default=f"""inference_memory_{round(time() )}.csv""" ,metadata={"help": "CSV filename used if saving memory results to csv."} ,) __lowerCAmelCase = field( default=f"""train_time_{round(time() )}.csv""" ,metadata={"help": "CSV filename used if saving time results to csv for training."} ,) __lowerCAmelCase = field( default=f"""train_memory_{round(time() )}.csv""" ,metadata={"help": "CSV filename used if saving memory results to csv for training."} ,) __lowerCAmelCase = field( default=f"""env_info_{round(time() )}.csv""" ,metadata={"help": "CSV filename used if saving environment information."} ,) __lowerCAmelCase = field( default=f"""log_{round(time() )}.csv""" ,metadata={"help": "Log filename used if print statements are saved in log."} ,) __lowerCAmelCase = field(default=3 ,metadata={"help": "Times an experiment will be run."} ) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={ "help": ( "Instead of loading the model as defined in `config.architectures` if exists, just load the pretrain" " model weights." ) } ,) def snake_case_ ( self ): warnings.warn( F"""The class {self.__class__} is deprecated. Hugging Face Benchmarking utils""" " are deprecated in general and it is advised to use external Benchmarking libraries " " to benchmark Transformer models." , snake_case_ , ) def snake_case_ ( self ): return json.dumps(dataclasses.asdict(self ) , indent=2 ) @property def snake_case_ ( self ): if len(self.models ) <= 0: raise ValueError( "Please make sure you provide at least one model name / model identifier, *e.g.* `--models" " bert-base-cased` or `args.models = [\'bert-base-cased\']." ) return self.models @property def snake_case_ ( self ): if not self.multi_process: return False elif self.is_tpu: logger.info("Multiprocessing is currently not possible on TPU." ) return False else: return True
237
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
0
"""simple docstring""" def _lowerCamelCase ( UpperCAmelCase_ : Optional[Any] ) -> Optional[Any]: """simple docstring""" A__ = [] A__ = set({"(", "[", "{"} ) A__ = set({")", "]", "}"} ) A__ = {"{": "}", "[": "]", "(": ")"} for i in range(len(lowercase_ ) ): if s[i] in open_brackets: stack.append(s[i] ) elif s[i] in closed_brackets and ( len(lowercase_ ) == 0 or (len(lowercase_ ) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(lowercase_ ) == 0 def _lowerCamelCase ( ) -> Any: """simple docstring""" A__ = input("Enter sequence of brackets: " ) if is_balanced(lowercase_ ): print(lowercase_, "is balanced" ) else: print(lowercase_, "is not balanced" ) if __name__ == "__main__": main()
104
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
0
"""simple docstring""" import math from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { '''facebook/data2vec-base-960h''': '''https://huggingface.co/facebook/data2vec-audio-base-960h/resolve/main/config.json''', # See all Data2VecAudio models at https://huggingface.co/models?filter=data2vec-audio } class lowerCamelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'data2vec-audio' def __init__(self , _lowerCamelCase=32 , _lowerCamelCase=768 , _lowerCamelCase=12 , _lowerCamelCase=12 , _lowerCamelCase=3072 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=0.0 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=0.02 , _lowerCamelCase=1e-5 , _lowerCamelCase="gelu" , _lowerCamelCase=(512, 512, 512, 512, 512, 512, 512) , _lowerCamelCase=(5, 2, 2, 2, 2, 2, 2) , _lowerCamelCase=(10, 3, 3, 3, 3, 2, 2) , _lowerCamelCase=False , _lowerCamelCase=16 , _lowerCamelCase=19 , _lowerCamelCase=5 , _lowerCamelCase=0.05 , _lowerCamelCase=10 , _lowerCamelCase=2 , _lowerCamelCase=0.0 , _lowerCamelCase=10 , _lowerCamelCase=0 , _lowerCamelCase="sum" , _lowerCamelCase=False , _lowerCamelCase=False , _lowerCamelCase=256 , _lowerCamelCase=(512, 512, 512, 512, 1500) , _lowerCamelCase=(5, 3, 3, 1, 1) , _lowerCamelCase=(1, 2, 3, 1, 1) , _lowerCamelCase=512 , _lowerCamelCase=0 , _lowerCamelCase=1 , _lowerCamelCase=2 , _lowerCamelCase=False , _lowerCamelCase=3 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=None , **_lowerCamelCase , ): """simple docstring""" super().__init__(**snake_case_ , pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ ) UpperCAmelCase__ : str = hidden_size UpperCAmelCase__ : Dict = feat_extract_activation UpperCAmelCase__ : Any = list(snake_case_ ) UpperCAmelCase__ : Dict = list(snake_case_ ) UpperCAmelCase__ : Optional[int] = list(snake_case_ ) UpperCAmelCase__ : List[Any] = conv_bias UpperCAmelCase__ : List[Any] = num_conv_pos_embeddings UpperCAmelCase__ : Any = num_conv_pos_embedding_groups UpperCAmelCase__ : int = conv_pos_kernel_size UpperCAmelCase__ : int = len(self.conv_dim ) UpperCAmelCase__ : Tuple = num_hidden_layers UpperCAmelCase__ : List[str] = intermediate_size UpperCAmelCase__ : int = hidden_act UpperCAmelCase__ : List[Any] = num_attention_heads UpperCAmelCase__ : Dict = hidden_dropout UpperCAmelCase__ : Optional[int] = attention_dropout UpperCAmelCase__ : Optional[int] = activation_dropout UpperCAmelCase__ : Union[str, Any] = feat_proj_dropout UpperCAmelCase__ : Any = final_dropout UpperCAmelCase__ : List[Any] = layerdrop UpperCAmelCase__ : Any = layer_norm_eps UpperCAmelCase__ : Union[str, Any] = initializer_range UpperCAmelCase__ : Optional[Any] = vocab_size UpperCAmelCase__ : Dict = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" F""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase__ : Dict = mask_time_prob UpperCAmelCase__ : List[str] = mask_time_length UpperCAmelCase__ : Union[str, Any] = mask_time_min_masks UpperCAmelCase__ : Optional[int] = mask_feature_prob UpperCAmelCase__ : List[Any] = mask_feature_length UpperCAmelCase__ : List[Any] = mask_feature_min_masks # ctc loss UpperCAmelCase__ : Tuple = ctc_loss_reduction UpperCAmelCase__ : str = ctc_zero_infinity # adapter UpperCAmelCase__ : Optional[int] = add_adapter UpperCAmelCase__ : List[str] = adapter_kernel_size UpperCAmelCase__ : str = adapter_stride UpperCAmelCase__ : Tuple = num_adapter_layers UpperCAmelCase__ : List[Any] = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase__ : Any = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase__ : List[Any] = list(snake_case_ ) UpperCAmelCase__ : Union[str, Any] = list(snake_case_ ) UpperCAmelCase__ : str = list(snake_case_ ) UpperCAmelCase__ : Any = xvector_output_dim @property def _a (self ): """simple docstring""" return math.prod(self.conv_stride )
182
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
0
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=30, lowerCamelCase=2, lowerCamelCase=3, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=32, lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=37, lowerCamelCase="gelu", lowerCamelCase=0.1, lowerCamelCase=0.1, lowerCamelCase=10, lowerCamelCase=0.0_2, lowerCamelCase=3, lowerCamelCase=0.6, lowerCamelCase=None, ) -> Optional[Any]: """simple docstring""" _lowercase : Any = parent _lowercase : Tuple = batch_size _lowercase : Union[str, Any] = image_size _lowercase : List[Any] = patch_size _lowercase : List[str] = num_channels _lowercase : Tuple = is_training _lowercase : str = use_labels _lowercase : Dict = hidden_size _lowercase : Union[str, Any] = num_hidden_layers _lowercase : Union[str, Any] = num_attention_heads _lowercase : Optional[Any] = intermediate_size _lowercase : List[str] = hidden_act _lowercase : str = hidden_dropout_prob _lowercase : List[Any] = attention_probs_dropout_prob _lowercase : Optional[Any] = type_sequence_label_size _lowercase : Dict = initializer_range _lowercase : Dict = mask_ratio _lowercase : int = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowercase : Tuple = (image_size // patch_size) ** 2 _lowercase : Dict = int(math.ceil((1 - mask_ratio) * (num_patches + 1))) def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) _lowercase : Tuple = None if self.use_labels: _lowercase : List[Any] = ids_tensor([self.batch_size], self.type_sequence_label_size) _lowercase : Any = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self) -> Any: """simple docstring""" return ViTMAEConfig( image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=snake_case_, initializer_range=self.initializer_range, mask_ratio=self.mask_ratio, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" _lowercase : str = ViTMAEModel(config=snake_case_) model.to(snake_case_) model.eval() _lowercase : Union[str, Any] = model(snake_case_) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Tuple: """simple docstring""" _lowercase : Union[str, Any] = ViTMAEForPreTraining(snake_case_) model.to(snake_case_) model.eval() _lowercase : str = model(snake_case_) _lowercase : Tuple = (self.image_size // self.patch_size) ** 2 _lowercase : Optional[Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape, (self.batch_size, num_patches, expected_num_channels)) # test greyscale images _lowercase : str = 1 _lowercase : Dict = ViTMAEForPreTraining(snake_case_) model.to(snake_case_) model.eval() _lowercase : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) _lowercase : Any = model(snake_case_) _lowercase : List[Any] = self.patch_size**2 self.parent.assertEqual(result.logits.shape, (self.batch_size, num_patches, expected_num_channels)) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = self.prepare_config_and_inputs() _lowercase , _lowercase , _lowercase : int = config_and_inputs _lowercase : Dict = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class _lowerCamelCase( __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, unittest.TestCase ): lowercase_ : Optional[Any] = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () lowercase_ : Tuple = {"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} lowercase_ : List[Any] = False lowercase_ : Optional[Any] = False lowercase_ : List[Any] = False lowercase_ : List[Any] = False def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Union[str, Any] = ViTMAEModelTester(self) _lowercase : List[str] = ConfigTester(self, config_class=snake_case_, has_text_modality=snake_case_, hidden_size=37) def UpperCamelCase ( self) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds') def UpperCamelCase ( self) -> Dict: """simple docstring""" pass def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase , _lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : str = model_class(snake_case_) self.assertIsInstance(model.get_input_embeddings(), (nn.Module)) _lowercase : List[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case_, nn.Linear)) def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase , _lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : Tuple = model_class(snake_case_) _lowercase : Optional[Any] = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase : Any = [*signature.parameters.keys()] _lowercase : Tuple = ['pixel_values'] self.assertListEqual(arg_names[:1], snake_case_) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_) def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*snake_case_) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[Any]: """simple docstring""" np.random.seed(2) _lowercase : Union[str, Any] = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2) _lowercase : List[str] = np.random.uniform(size=(self.model_tester.batch_size, num_patches)) _lowercase : Optional[Any] = torch.from_numpy(snake_case_) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowercase : Tuple = pt_noise super().check_pt_tf_models(snake_case_, snake_case_, snake_case_) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase , _lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : List[str] = model_class(snake_case_) model.to(snake_case_) model.eval() # make random mask reproducible torch.manual_seed(2) with torch.no_grad(): _lowercase : Dict = model(**self._prepare_for_class(snake_case_, snake_case_)) _lowercase : Dict = outputs[0].cpu().numpy() _lowercase : Optional[Any] = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_) _lowercase : Union[str, Any] = model_class.from_pretrained(snake_case_) model.to(snake_case_) # make random mask reproducible torch.manual_seed(2) with torch.no_grad(): _lowercase : Any = model(**self._prepare_for_class(snake_case_, snake_case_)) # Make sure we don't have nans _lowercase : List[Any] = after_outputs[0].cpu().numpy() _lowercase : List[Any] = 0 _lowercase : Dict = np.amax(np.abs(out_a - out_a)) self.assertLessEqual(snake_case_, 1E-5) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.') def UpperCamelCase ( self) -> List[Any]: """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.') def UpperCamelCase ( self) -> List[Any]: """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.') def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load') def UpperCamelCase ( self) -> Dict: """simple docstring""" pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.') def UpperCamelCase ( self) -> Dict: """simple docstring""" pass @slow def UpperCamelCase ( self) -> Tuple: """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : str = ViTMAEModel.from_pretrained(snake_case_) self.assertIsNotNone(snake_case_) def UpperCamelCase_( ) -> Any: _lowercase : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class _lowerCamelCase( unittest.TestCase ): @cached_property def UpperCamelCase ( self) -> List[Any]: """simple docstring""" return ViTImageProcessor.from_pretrained('facebook/vit-mae-base') if is_vision_available() else None @slow def UpperCamelCase ( self) -> int: """simple docstring""" np.random.seed(2) _lowercase : int = ViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base').to(snake_case_) _lowercase : Any = self.default_image_processor _lowercase : int = prepare_img() _lowercase : Union[str, Any] = image_processor(images=snake_case_, return_tensors='pt').to(snake_case_) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowercase : List[str] = ViTMAEConfig() _lowercase : List[Any] = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2) _lowercase : int = np.random.uniform(size=(1, num_patches)) # forward pass with torch.no_grad(): _lowercase : List[str] = model(**snake_case_, noise=torch.from_numpy(snake_case_).to(device=snake_case_)) # verify the logits _lowercase : str = torch.Size((1, 1_96, 7_68)) self.assertEqual(outputs.logits.shape, snake_case_) _lowercase : Dict = torch.tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]]) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3], expected_slice.to(snake_case_), atol=1E-4))
89
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
0
from math import ceil, sqrt def __UpperCAmelCase( lowercase_ = 1_00_00_00 ): _lowerCamelCase : Any = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: _lowerCamelCase : str = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: _lowerCamelCase : Optional[int] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(F'''{solution() = }''')
114
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
0
from numpy import exp, pi, sqrt def UpperCamelCase ( __lowercase : Union[str, Any] ,__lowercase : float = 0.0 ,__lowercase : float = 1.0 ): '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
558
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
0
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets UpperCamelCase_ = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' UpperCamelCase_ = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' UpperCamelCase_ = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ (datasets.Metric ): def __UpperCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/google-research/google-research/tree/master/rouge"""] , reference_urls=[ """https://en.wikipedia.org/wiki/ROUGE_(metric)""", """https://github.com/google-research/google-research/tree/master/rouge""", ] , ) def __UpperCamelCase ( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: _lowerCAmelCase : Any = ["""rouge1""", """rouge2""", """rougeL""", """rougeLsum"""] _lowerCAmelCase : str = rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: _lowerCAmelCase : Optional[Any] = scoring.BootstrapAggregator() else: _lowerCAmelCase : Union[str, Any] = [] for ref, pred in zip(snake_case_ , snake_case_ ): _lowerCAmelCase : int = scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: _lowerCAmelCase : Optional[Any] = aggregator.aggregate() else: _lowerCAmelCase : int = {} for key in scores[0]: _lowerCAmelCase : Dict = [score[key] for score in scores] return result
384
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
0
"""simple docstring""" from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch('''socket.socket''' ) @patch('''builtins.open''' ) def UpperCAmelCase ( snake_case : int , snake_case : List[str] ): _lowerCAmelCase:Union[str, Any] = Mock() _lowerCAmelCase:List[str] = conn, Mock() _lowerCAmelCase:Union[str, Any] = iter([1, None] ) _lowerCAmelCase:Optional[Any] = lambda snake_case : next(lowercase_ ) # ===== invoke ===== send_file(filename='''mytext.txt''' , testing=lowercase_ ) # ===== ensurance ===== sock.assert_called_once() sock.return_value.bind.assert_called_once() sock.return_value.listen.assert_called_once() sock.return_value.accept.assert_called_once() conn.recv.assert_called_once() file.return_value.__enter__.assert_called_once() file.return_value.__enter__.return_value.read.assert_called() conn.send.assert_called_once() conn.close.assert_called_once() sock.return_value.shutdown.assert_called_once() sock.return_value.close.assert_called_once()
227
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed __magic_name__ : Dict = logging.getLogger(__name__) def lowercase__ ( _UpperCamelCase=2 , _UpperCamelCase=3 , _UpperCamelCase=16 , _UpperCamelCase = 10 , _UpperCamelCase = 2) -> Dict: """simple docstring""" def get_dataset(_UpperCamelCase): UpperCamelCase = torch.randn(batch_size * n_batches , 1) return TensorDataset(lowercase_ , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1)) UpperCamelCase = get_dataset(lowercase_) UpperCamelCase = get_dataset(lowercase_) UpperCamelCase = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4) UpperCamelCase = DataLoader(lowercase_ , shuffle=lowercase_ , batch_size=lowercase_ , num_workers=4) return (train_dataloader, valid_dataloader) def lowercase__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=None) -> List[str]: """simple docstring""" UpperCamelCase = [] for epoch in range(lowercase_): # Train quickly model.train() for batch in dataloader: UpperCamelCase , UpperCamelCase = batch UpperCamelCase = model(lowercase_) UpperCamelCase = torch.nn.functional.mse_loss(lowercase_ , lowercase_) accelerator.backward(lowercase_) optimizer.step() optimizer.zero_grad() rands.append(random.random()) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class A__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] ): """simple docstring""" super().__init__() UpperCamelCase = nn.Parameter(torch.randn(1 ) ) UpperCamelCase = nn.Parameter(torch.randn(1 ) ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" return x * self.a + self.b class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase = DummyModel() UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) UpperCamelCase , UpperCamelCase = dummy_dataloaders() UpperCamelCase = ProjectConfiguration(total_limit=1 , project_dir=snake_case_ , automatic_checkpoint_naming=snake_case_ ) # Train baseline UpperCamelCase = Accelerator(project_config=snake_case_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase = DummyModel() UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) UpperCamelCase , UpperCamelCase = dummy_dataloaders() # Train baseline UpperCamelCase = Accelerator() UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Save initial UpperCamelCase = os.path.join(snake_case_ , 'initial' ) accelerator.save_state(snake_case_ ) ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() UpperCamelCase = train(3 , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() # Train partially set_seed(42 ) UpperCamelCase = DummyModel() UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) UpperCamelCase , UpperCamelCase = dummy_dataloaders() UpperCamelCase = Accelerator() UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) accelerator.load_state(snake_case_ ) ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) UpperCamelCase = train(2 , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Save everything UpperCamelCase = os.path.join(snake_case_ , 'checkpoint' ) accelerator.save_state(snake_case_ ) # Load everything back in and make sure all states work accelerator.load_state(snake_case_ ) test_rands += train(1 , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) def _SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase = DummyModel() UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) UpperCamelCase , UpperCamelCase = dummy_dataloaders() UpperCamelCase = ProjectConfiguration(automatic_checkpoint_naming=snake_case_ ) # Train baseline UpperCamelCase = Accelerator(project_dir=snake_case_ , project_config=snake_case_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Save initial accelerator.save_state() ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() UpperCamelCase = train(3 , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() # Train partially set_seed(42 ) UpperCamelCase = DummyModel() UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) UpperCamelCase , UpperCamelCase = dummy_dataloaders() UpperCamelCase = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=snake_case_ ) UpperCamelCase = Accelerator(project_dir=snake_case_ , project_config=snake_case_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) accelerator.load_state(os.path.join(snake_case_ , 'checkpoints' , 'checkpoint_0' ) ) ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) UpperCamelCase = train(2 , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(snake_case_ , 'checkpoints' , 'checkpoint_1' ) ) test_rands += train(1 , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) ((UpperCamelCase) , (UpperCamelCase)) = model.a.item(), model.b.item() UpperCamelCase = optimizer.state_dict() self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" UpperCamelCase = torch.tensor([1, 2, 3] ) UpperCamelCase = torch.tensor([2, 3, 4] ) UpperCamelCase = DummyModel() UpperCamelCase = torch.optim.Adam(net.parameters() ) UpperCamelCase = Accelerator() with self.assertRaises(snake_case_ ) as ve: accelerator.register_for_checkpointing(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) UpperCamelCase = str(ve.exception ) self.assertTrue('Item at index 0' in message ) self.assertTrue('Item at index 1' in message ) self.assertFalse('Item at index 2' in message ) self.assertFalse('Item at index 3' in message ) def _SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase = DummyModel() UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) UpperCamelCase = torch.optim.lr_scheduler.StepLR(snake_case_ , step_size=1 , gamma=0.9_9 ) UpperCamelCase , UpperCamelCase = dummy_dataloaders() UpperCamelCase = ProjectConfiguration(automatic_checkpoint_naming=snake_case_ ) # Train baseline UpperCamelCase = Accelerator(project_dir=snake_case_ , project_config=snake_case_ ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Save initial accelerator.save_state() UpperCamelCase = scheduler.state_dict() train(3 , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.assertNotEqual(snake_case_ , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(snake_case_ , 'checkpoints' , 'checkpoint_0' ) ) self.assertEqual(snake_case_ , scheduler.state_dict() ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCamelCase = DummyModel() UpperCamelCase = ProjectConfiguration(automatic_checkpoint_naming=snake_case_ , total_limit=2 ) # Train baseline UpperCamelCase = Accelerator(project_dir=snake_case_ , project_config=snake_case_ ) UpperCamelCase = accelerator.prepare(snake_case_ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(snake_case_ , 'checkpoints' , 'checkpoint_0' ) ) ) self.assertTrue(os.path.exists(os.path.join(snake_case_ , 'checkpoints' , 'checkpoint_9' ) ) ) self.assertTrue(os.path.exists(os.path.join(snake_case_ , 'checkpoints' , 'checkpoint_10' ) ) ) @require_cuda def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" UpperCamelCase = ['torchrun', f'--nproc_per_node={torch.cuda.device_count()}', inspect.getfile(self.__class__ )] execute_subprocess_async(snake_case_ , env=os.environ.copy() ) if __name__ == "__main__": __magic_name__ : List[str] = '''/tmp/accelerate/state_checkpointing''' __magic_name__ : str = DummyModel() __magic_name__ : Optional[int] = torch.optim.Adam(params=model.parameters(), lr=1e-3) __magic_name__ : Tuple = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) __magic_name__ : Union[str, Any] = dummy_dataloaders() __magic_name__ : List[str] = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline __magic_name__ : Optional[int] = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='''no''') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) __magic_name__ : Dict = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) __magic_name__ : Tuple = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: __magic_name__ : Optional[Any] = group['''params'''][0].device break assert param_device.type == accelerator.device.type __magic_name__ : List[Any] = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''cpu''') for group in optimizer.param_groups: __magic_name__ : Any = group['''params'''][0].device break assert ( param_device.type == torch.device('''cpu''').type ), F"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''on_device''') for group in optimizer.param_groups: __magic_name__ : int = group['''params'''][0].device break assert ( param_device.type == accelerator.device.type ), F"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='''Unsupported optimizer map location passed'''): accelerator.load_state(os.path.join(savedir, '''checkpoints''', '''checkpoint_0'''), map_location='''invalid''') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
280
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase :Any = logging.get_logger(__name__) lowerCamelCase :Dict = { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/config.json''', # See all XGLM models at https://huggingface.co/models?filter=xglm } class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : Optional[int] = 'xglm' __SCREAMING_SNAKE_CASE : Dict = ['past_key_values'] __SCREAMING_SNAKE_CASE : Tuple = { 'num_attention_heads': 'attention_heads', 'hidden_size': 'd_model', 'num_hidden_layers': 'num_layers', } def __init__(self , lowercase=256008 , lowercase=2048 , lowercase=1024 , lowercase=4096 , lowercase=24 , lowercase=16 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=2 , lowercase=1 , lowercase=0 , lowercase=2 , **lowercase , ): A_ : str = vocab_size A_ : str = max_position_embeddings A_ : List[Any] = d_model A_ : Tuple = ffn_dim A_ : List[Any] = num_layers A_ : Optional[int] = attention_heads A_ : Optional[int] = activation_function A_ : List[str] = dropout A_ : Dict = attention_dropout A_ : int = activation_dropout A_ : Optional[int] = layerdrop A_ : Tuple = init_std A_ : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True A_ : str = use_cache super().__init__( pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , **snake_case_ , )
667
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
0
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase : Tuple =["image_processor", "tokenizer"] _UpperCAmelCase : Any ="BlipImageProcessor" _UpperCAmelCase : Union[str, Any] ="AutoTokenizer" def __init__( self : Optional[int] , lowerCAmelCase : List[Any] , lowerCAmelCase : List[str] ): A_ = False super().__init__(snake_case_ , snake_case_ ) A_ = self.image_processor def __call__( self : Tuple , lowerCAmelCase : Tuple = None , lowerCAmelCase : List[str] = None , lowerCAmelCase : Any = True , lowerCAmelCase : Any = False , lowerCAmelCase : Union[str, Any] = None , lowerCAmelCase : Optional[Any] = None , lowerCAmelCase : Tuple = 0 , lowerCAmelCase : str = None , lowerCAmelCase : str = None , lowerCAmelCase : str = False , lowerCAmelCase : Union[str, Any] = False , lowerCAmelCase : List[str] = False , lowerCAmelCase : str = False , lowerCAmelCase : Optional[int] = False , lowerCAmelCase : Dict = True , lowerCAmelCase : Union[str, Any] = None , **lowerCAmelCase : Union[str, Any] , ): if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: A_ = self.tokenizer A_ = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) return text_encoding # add pixel_values A_ = self.image_processor(snake_case_ , return_tensors=snake_case_ ) if text is not None: A_ = self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) else: A_ = None if text_encoding is not None: encoding_image_processor.update(snake_case_ ) return encoding_image_processor def _UpperCAmelCase ( self : Any , *lowerCAmelCase : int , **lowerCAmelCase : int ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _UpperCAmelCase ( self : str , *lowerCAmelCase : Dict , **lowerCAmelCase : int ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _UpperCAmelCase ( self : int ): A_ = self.tokenizer.model_input_names A_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
452
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
0
"""simple docstring""" def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ ) -> int: if n == 1 or not isinstance(lowercase_, lowercase_ ): return 0 elif n == 2: return 1 else: a_ : List[Any] = [0, 1] for i in range(2, n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ ) -> int: a_ : Dict = 0 a_ : Union[str, Any] = 2 while digits < n: index += 1 a_ : int = len(str(fibonacci(lowercase_ ) ) ) return index def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ = 1_000 ) -> int: return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
237
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
0
"""simple docstring""" import math def _lowerCamelCase ( UpperCAmelCase_ : int ) -> bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5, int(math.sqrt(lowercase_ ) + 1 ), 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _lowerCamelCase ( UpperCAmelCase_ : float = 0.1 ) -> int: """simple docstring""" A__ = 3 A__ = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1, (j + 2) * (j + 2), j + 1 ): primes += is_prime(lowercase_ ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
104
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
"""simple docstring""" from __future__ import annotations def a__ ( lowerCAmelCase ) -> float: UpperCAmelCase__ : Dict = 0.00 UpperCAmelCase__ : Any = 0 for resistor in resistors: if resistor <= 0: UpperCAmelCase__ : str = F"""Resistor at index {index} has a negative or zero value!""" raise ValueError(lowercase_ ) first_sum += 1 / float(lowercase_ ) index += 1 return 1 / first_sum def a__ ( lowerCAmelCase ) -> float: UpperCAmelCase__ : Tuple = 0.00 UpperCAmelCase__ : List[str] = 0 for resistor in resistors: sum_r += resistor if resistor < 0: UpperCAmelCase__ : Tuple = F"""Resistor at index {index} has a negative value!""" raise ValueError(lowercase_ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
182
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : str = { '''shi-labs/nat-mini-in1k-224''': '''https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json''', # See all Nat models at https://huggingface.co/models?filter=nat } class _lowerCamelCase( __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE ): lowercase_ : Tuple = """nat""" lowercase_ : Any = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self, lowerCamelCase=4, lowerCamelCase=3, lowerCamelCase=64, lowerCamelCase=[3, 4, 6, 5], lowerCamelCase=[2, 4, 8, 16], lowerCamelCase=7, lowerCamelCase=3.0, lowerCamelCase=True, lowerCamelCase=0.0, lowerCamelCase=0.0, lowerCamelCase=0.1, lowerCamelCase="gelu", lowerCamelCase=0.0_2, lowerCamelCase=1E-5, lowerCamelCase=0.0, lowerCamelCase=None, lowerCamelCase=None, **lowerCamelCase, ) -> Dict: """simple docstring""" super().__init__(**snake_case_) _lowercase : Tuple = patch_size _lowercase : Tuple = num_channels _lowercase : List[Any] = embed_dim _lowercase : str = depths _lowercase : List[str] = len(snake_case_) _lowercase : str = num_heads _lowercase : List[str] = kernel_size _lowercase : List[str] = mlp_ratio _lowercase : List[Any] = qkv_bias _lowercase : Optional[int] = hidden_dropout_prob _lowercase : Any = attention_probs_dropout_prob _lowercase : Optional[Any] = drop_path_rate _lowercase : Optional[Any] = hidden_act _lowercase : Optional[Any] = layer_norm_eps _lowercase : Optional[Any] = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _lowercase : Optional[Any] = int(embed_dim * 2 ** (len(snake_case_) - 1)) _lowercase : str = layer_scale_init_value _lowercase : Optional[Any] = ['stem'] + [F'''stage{idx}''' for idx in range(1, len(snake_case_) + 1)] _lowercase , _lowercase : Union[str, Any] = get_aligned_output_features_output_indices( out_features=snake_case_, out_indices=snake_case_, stage_names=self.stage_names)
89
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __A ( unittest.TestCase ): """simple docstring""" def __init__( self , a__ , a__=7 , a__=3 , a__=18 , a__=30 , a__=400 , a__=True , a__=None , a__=True , a__=None , a__=True , a__=[0.5, 0.5, 0.5] , a__=[0.5, 0.5, 0.5] , ): """simple docstring""" _lowerCamelCase : Dict = size if size is not None else {'''shortest_edge''': 18} _lowerCamelCase : int = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _lowerCamelCase : List[Any] = parent _lowerCamelCase : Tuple = batch_size _lowerCamelCase : List[str] = num_channels _lowerCamelCase : List[str] = image_size _lowerCamelCase : List[Any] = min_resolution _lowerCamelCase : List[str] = max_resolution _lowerCamelCase : Dict = do_resize _lowerCamelCase : Any = size _lowerCamelCase : Tuple = do_center_crop _lowerCamelCase : Optional[int] = crop_size _lowerCamelCase : Optional[int] = do_normalize _lowerCamelCase : List[Any] = image_mean _lowerCamelCase : Optional[Any] = image_std def __snake_case ( self): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __A ( __SCREAMING_SNAKE_CASE ,unittest.TestCase ): """simple docstring""" UpperCAmelCase__ = LevitImageProcessor if is_vision_available() else None def __snake_case ( self): """simple docstring""" _lowerCamelCase : Union[str, Any] = LevitImageProcessingTester(self) @property def __snake_case ( self): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self): """simple docstring""" _lowerCamelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(snake_case_ , '''image_mean''')) self.assertTrue(hasattr(snake_case_ , '''image_std''')) self.assertTrue(hasattr(snake_case_ , '''do_normalize''')) self.assertTrue(hasattr(snake_case_ , '''do_resize''')) self.assertTrue(hasattr(snake_case_ , '''do_center_crop''')) self.assertTrue(hasattr(snake_case_ , '''size''')) def __snake_case ( self): """simple docstring""" _lowerCamelCase : Tuple = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 18}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) _lowerCamelCase : int = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def __snake_case ( self): """simple docstring""" pass def __snake_case ( self): """simple docstring""" _lowerCamelCase : Dict = self.image_processing_class(**self.image_processor_dict) # create random PIL images _lowerCamelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_) for image in image_inputs: self.assertIsInstance(snake_case_ , Image.Image) # Test not batched input _lowerCamelCase : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _lowerCamelCase : int = image_processing(snake_case_ , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def __snake_case ( self): """simple docstring""" _lowerCamelCase : Tuple = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _lowerCamelCase : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ , numpify=snake_case_) for image in image_inputs: self.assertIsInstance(snake_case_ , np.ndarray) # Test not batched input _lowerCamelCase : Any = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _lowerCamelCase : int = image_processing(snake_case_ , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def __snake_case ( self): """simple docstring""" _lowerCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _lowerCamelCase : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case_ , torchify=snake_case_) for image in image_inputs: self.assertIsInstance(snake_case_ , torch.Tensor) # Test not batched input _lowerCamelCase : Tuple = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _lowerCamelCase : Optional[int] = image_processing(snake_case_ , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
114
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
0
import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ = DanceDiffusionPipeline lowerCamelCase_ = UNCONDITIONAL_AUDIO_GENERATION_PARAMS lowerCamelCase_ = PipelineTesterMixin.required_optional_params - { '''callback''', '''latents''', '''callback_steps''', '''output_type''', '''num_images_per_prompt''', } lowerCamelCase_ = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS lowerCamelCase_ = False lowerCamelCase_ = False def lowerCAmelCase_ ( self ): """simple docstring""" torch.manual_seed(0 ) A_ : int = UNetaDModel( block_out_channels=(3_2, 3_2, 6_4) , extra_in_channels=1_6 , sample_size=5_1_2 , sample_rate=1_6_0_0_0 , in_channels=2 , out_channels=2 , flip_sin_to_cos=snake_case_ , use_timestep_embedding=snake_case_ , time_embedding_type='fourier' , mid_block_type='UNetMidBlock1D' , down_block_types=('DownBlock1DNoSkip', 'DownBlock1D', 'AttnDownBlock1D') , up_block_types=('AttnUpBlock1D', 'UpBlock1D', 'UpBlock1DNoSkip') , ) A_ : Tuple = IPNDMScheduler() A_ : str = { 'unet': unet, 'scheduler': scheduler, } return components def lowerCAmelCase_ ( self , lowercase , lowercase=0 ): """simple docstring""" if str(snake_case_ ).startswith('mps' ): A_ : str = torch.manual_seed(snake_case_ ) else: A_ : Any = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) A_ : str = { 'batch_size': 1, 'generator': generator, 'num_inference_steps': 4, } return inputs def lowerCAmelCase_ ( self ): """simple docstring""" A_ : int = 'cpu' # ensure determinism for the device-dependent torch.Generator A_ : List[str] = self.get_dummy_components() A_ : Tuple = DanceDiffusionPipeline(**snake_case_ ) A_ : Dict = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Tuple = self.get_dummy_inputs(snake_case_ ) A_ : Optional[int] = pipe(**snake_case_ ) A_ : Optional[Any] = output.audios A_ : Tuple = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) A_ : Tuple = np.array([-0.7265, 1.0000, -0.8388, 0.1175, 0.9498, -1.0000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def lowerCAmelCase_ ( self ): """simple docstring""" return super().test_save_load_local() @skip_mps def lowerCAmelCase_ ( self ): """simple docstring""" return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) @skip_mps def lowerCAmelCase_ ( self ): """simple docstring""" return super().test_save_load_optional_components() @skip_mps def lowerCAmelCase_ ( self ): """simple docstring""" return super().test_attention_slicing_forward_pass() def lowerCAmelCase_ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase_ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Union[str, Any] = torch_device A_ : Any = DanceDiffusionPipeline.from_pretrained('harmonai/maestro-150k' ) A_ : List[str] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) A_ : Union[str, Any] = torch.manual_seed(0 ) A_ : Union[str, Any] = pipe(generator=snake_case_ , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) A_ : List[str] = output.audios A_ : int = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) A_ : int = np.array([-0.0192, -0.0231, -0.0318, -0.0059, 0.0002, -0.0020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase_ ( self ): """simple docstring""" A_ : Dict = torch_device A_ : Dict = DanceDiffusionPipeline.from_pretrained('harmonai/maestro-150k' , torch_dtype=torch.floataa ) A_ : str = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) A_ : List[Any] = torch.manual_seed(0 ) A_ : Dict = pipe(generator=snake_case_ , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) A_ : List[str] = output.audios A_ : Union[str, Any] = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) A_ : Optional[Any] = np.array([-0.0367, -0.0488, -0.0771, -0.0525, -0.0444, -0.0341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2
558
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
0
'''simple docstring''' from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo UpperCamelCase_ = '''\ @misc{wu2016googles, title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes and Jeffrey Dean}, year={2016}, eprint={1609.08144}, archivePrefix={arXiv}, primaryClass={cs.CL} } ''' UpperCamelCase_ = '''\ The BLEU score has some undesirable properties when used for single sentences, as it was designed to be a corpus measure. We therefore use a slightly different score for our RL experiments which we call the \'GLEU score\'. For the GLEU score, we record all sub-sequences of 1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then compute a recall, which is the ratio of the number of matching n-grams to the number of total n-grams in the target (ground truth) sequence, and a precision, which is the ratio of the number of matching n-grams to the number of total n-grams in the generated output sequence. Then GLEU score is simply the minimum of recall and precision. This GLEU score\'s range is always between 0 (no matches) and 1 (all match) and it is symmetrical when switching output and target. According to our experiments, GLEU score correlates quite well with the BLEU metric on a corpus level but does not have its drawbacks for our per sentence reward objective. ''' UpperCamelCase_ = '''\ Computes corpus-level Google BLEU (GLEU) score of translated segments against one or more references. Instead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching tokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values. Args: predictions (list of str): list of translations to score. Each translation should be tokenized into a list of tokens. references (list of list of str): list of lists of references for each translation. Each reference should be tokenized into a list of tokens. min_len (int): The minimum order of n-gram this function should extract. Defaults to 1. max_len (int): The maximum order of n-gram this function should extract. Defaults to 4. Returns: \'google_bleu\': google_bleu score Examples: Example 1: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.44 Example 2: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.61 Example 3: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2) >>> print(round(results["google_bleu"], 2)) 0.53 Example 4: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6) >>> print(round(results["google_bleu"], 2)) 0.4 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ (datasets.Metric ): def __UpperCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" , id="""token""" ) , id="""sequence""" ), """references""": datasets.Sequence( datasets.Sequence(datasets.Value("""string""" , id="""token""" ) , id="""sequence""" ) , id="""references""" ), } ) , ) def __UpperCamelCase ( self , snake_case_ , snake_case_ , snake_case_ = 1 , snake_case_ = 4 , ): return { "google_bleu": gleu_score.corpus_gleu( list_of_references=snake_case_ , hypotheses=snake_case_ , min_len=snake_case_ , max_len=snake_case_ ) }
384
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
0
"""simple docstring""" import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class a__ ( __SCREAMING_SNAKE_CASE ): snake_case__ = 0 snake_case__ = False snake_case__ = 3.0 class a__ ( unittest.TestCase ): def __UpperCamelCase ( self : Optional[int]) -> Dict: """simple docstring""" self.assertDictEqual(MockClass().to_kwargs() ,{}) self.assertDictEqual(MockClass(a=2).to_kwargs() ,{'''a''': 2}) self.assertDictEqual(MockClass(a=2 ,b=snake_case_).to_kwargs() ,{'''a''': 2, '''b''': True}) self.assertDictEqual(MockClass(a=2 ,c=2.25).to_kwargs() ,{'''a''': 2, '''c''': 2.25}) @require_cuda def __UpperCamelCase ( self : List[str]) -> str: """simple docstring""" _lowerCAmelCase:int = GradScalerKwargs(init_scale=1024 ,growth_factor=2) AcceleratorState._reset_state() _lowerCAmelCase:List[str] = Accelerator(mixed_precision='''fp16''' ,kwargs_handlers=[scaler_handler]) print(accelerator.use_fpaa) _lowerCAmelCase:Any = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale ,1024.0) self.assertEqual(scaler._growth_factor ,2.0) # Check the other values are at the default self.assertEqual(scaler._backoff_factor ,0.5) self.assertEqual(scaler._growth_interval ,2000) self.assertEqual(scaler._enabled ,snake_case_) @require_multi_gpu def __UpperCamelCase ( self : int) -> str: """simple docstring""" _lowerCAmelCase:Optional[Any] = ['''torchrun''', F'--nproc_per_node={torch.cuda.device_count()}', inspect.getfile(self.__class__)] execute_subprocess_async(snake_case_ ,env=os.environ.copy()) if __name__ == "__main__": UpperCamelCase__ = DistributedDataParallelKwargs(bucket_cap_mb=1_5, find_unused_parameters=True) UpperCamelCase__ = Accelerator(kwargs_handlers=[ddp_scaler]) UpperCamelCase__ = torch.nn.Linear(1_0_0, 2_0_0) UpperCamelCase__ = accelerator.prepare(model) # Check the values changed in kwargs UpperCamelCase__ = '''''' UpperCamelCase__ = model.bucket_bytes_cap // (1_0_2_4 * 1_0_2_4) if observed_bucket_cap_map != 1_5: error_msg += F"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += F"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += F"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += F"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += F"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
227
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
from argparse import ArgumentParser from . import BaseTransformersCLICommand def lowercase__ ( _UpperCamelCase) -> Tuple: """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' @staticmethod def _SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE : List[Any] ): """simple docstring""" UpperCamelCase = parser.add_parser('download' ) download_parser.add_argument( '--cache-dir' , type=snake_case_ , default=snake_case_ , help='Path to location to store the models' ) download_parser.add_argument( '--force' , action='store_true' , help='Force the model to be download even if already in cache-dir' ) download_parser.add_argument( '--trust-remote-code' , action='store_true' , help='Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine' , ) download_parser.add_argument('model' , type=snake_case_ , help='Name of the model to download' ) download_parser.set_defaults(func=snake_case_ ) def __init__( self : str , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Any ): """simple docstring""" UpperCamelCase = model UpperCamelCase = cache UpperCamelCase = force UpperCamelCase = trust_remote_code def _SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
280
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase :List[Any] = {'''configuration_timm_backbone''': ['''TimmBackboneConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase :List[Any] = ['''TimmBackbone'''] if TYPE_CHECKING: from .configuration_timm_backbone import TimmBackboneConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timm_backbone import TimmBackbone else: import sys lowerCamelCase :Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
667
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : str = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Optional[Any] = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
452
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class snake_case_ ( __SCREAMING_SNAKE_CASE ,unittest.TestCase ): __lowerCAmelCase = KandinskyVaaPipeline __lowerCAmelCase = [ "image_embeds", "negative_image_embeds", ] __lowerCAmelCase = ["image_embeds", "negative_image_embeds"] __lowerCAmelCase = [ "generator", "height", "width", "latents", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] __lowerCAmelCase = False @property def snake_case_ ( self ): return 3_2 @property def snake_case_ ( self ): return 3_2 @property def snake_case_ ( self ): return self.time_input_dim @property def snake_case_ ( self ): return self.time_input_dim * 4 @property def snake_case_ ( self ): return 1_0_0 @property def snake_case_ ( self ): torch.manual_seed(0 ) a_ : int = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } a_ : int = UNetaDConditionModel(**snake_case_ ) return model @property def snake_case_ ( self ): return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def snake_case_ ( self ): torch.manual_seed(0 ) a_ : int = VQModel(**self.dummy_movq_kwargs ) return model def snake_case_ ( self ): a_ : Optional[Any] = self.dummy_unet a_ : Tuple = self.dummy_movq a_ : Optional[Any] = DDIMScheduler( num_train_timesteps=1_0_0_0 , beta_schedule="linear" , beta_start=0.00_085 , beta_end=0.012 , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , steps_offset=1 , prediction_type="epsilon" , thresholding=snake_case_ , ) a_ : int = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def snake_case_ ( self , a_ , a_=0 ): a_ : int = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) a_ : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( snake_case_ ) if str(snake_case_ ).startswith("mps" ): a_ : Dict = torch.manual_seed(snake_case_ ) else: a_ : Dict = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) a_ : List[str] = { "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 6_4, "width": 6_4, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def snake_case_ ( self ): a_ : List[Any] = "cpu" a_ : Union[str, Any] = self.get_dummy_components() a_ : Any = self.pipeline_class(**snake_case_ ) a_ : Optional[int] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) a_ : str = pipe(**self.get_dummy_inputs(snake_case_ ) ) a_ : str = output.images a_ : Union[str, Any] = pipe( **self.get_dummy_inputs(snake_case_ ) , return_dict=snake_case_ , )[0] a_ : List[str] = image[0, -3:, -3:, -1] a_ : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) a_ : Union[str, Any] = np.array( [0.6_237_976, 1.0, 0.36_441_332, 1.0, 0.70_639_634, 0.29_877_186, 0.85_652_125, 0.5_216_843, 0.54_454_046] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class snake_case_ ( unittest.TestCase ): def snake_case_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case_ ( self ): a_ : str = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy" ) a_ : Dict = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(snake_case_ ) a_ : Any = KandinskyVaaPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) a_ : Optional[Any] = pipeline.to(snake_case_ ) pipeline.set_progress_bar_config(disable=snake_case_ ) a_ : Optional[Any] = "red cat, 4k photo" a_ : Dict = torch.Generator(device="cuda" ).manual_seed(0 ) a_ , a_ : Tuple = pipe_prior( snake_case_ , generator=snake_case_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() a_ : Union[str, Any] = torch.Generator(device="cuda" ).manual_seed(0 ) a_ : str = pipeline( image_embeds=snake_case_ , negative_image_embeds=snake_case_ , generator=snake_case_ , num_inference_steps=1_0_0 , output_type="np" , ) a_ : Union[str, Any] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert_mean_pixel_difference(snake_case_ , snake_case_ )
237
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_text_model' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = { 'hidden_size': 'hidden_size', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , snake_case_=5_02_44 , snake_case_=7_68 , snake_case_=64 , snake_case_=20_48 , snake_case_=12 , snake_case_=12 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gelu_new" , snake_case_=0 , snake_case_=False , snake_case_=0 , snake_case_=1 , snake_case_=False , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =hidden_size lowercase =d_kv lowercase =d_ff lowercase =num_layers lowercase =num_heads lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =dropout_rate lowercase =layer_norm_epsilon lowercase =initializer_factor lowercase =use_cache lowercase =eos_token_id lowercase =decoder_start_token_id # for backwards compatibility lowercase =dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct_vision_model' def __init__( self , snake_case_=7_68 , snake_case_=7_68 , snake_case_=20_48 , snake_case_=64 , snake_case_=12 , snake_case_=12 , snake_case_="gelu_new" , snake_case_=1E-6 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-10 , snake_case_=1.0 , snake_case_=40_96 , snake_case_=32 , snake_case_=1_28 , **snake_case_ , ): super().__init__(**snake_case_ ) lowercase =hidden_size lowercase =patch_embed_hidden_size lowercase =d_ff lowercase =dropout_rate lowercase =num_hidden_layers lowercase =num_attention_heads lowercase =initializer_range lowercase =initializer_factor lowercase =attention_dropout lowercase =layer_norm_eps lowercase =dense_act_fn lowercase =seq_len lowercase =relative_attention_num_buckets lowercase =relative_attention_max_distance lowercase =d_kv @classmethod def _A( cls , snake_case_ , **snake_case_ ): cls._set_token_in_kwargs(snake_case_ ) lowercase , lowercase =cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('''model_type''' ) == "pix2struct": lowercase =config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'pix2struct' UpperCamelCase__ = True def __init__( self , snake_case_=None , snake_case_=None , snake_case_=1.0 , snake_case_=0.02 , snake_case_=False , snake_case_=False , snake_case_=True , **snake_case_ , ): super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: lowercase ={} logger.info('''text_config is None. Initializing the Pix2StructTextConfig with default values.''' ) if vision_config is None: lowercase ={} logger.info('''vision_config is None. Initializing the Pix2StructVisionConfig with default values.''' ) lowercase =PixaStructTextConfig(**snake_case_ ) lowercase =PixaStructVisionConfig(**snake_case_ ) lowercase =self.text_config.decoder_start_token_id lowercase =self.text_config.pad_token_id lowercase =self.text_config.eos_token_id lowercase =initializer_factor lowercase =initializer_range lowercase =self.initializer_range lowercase =self.initializer_range lowercase =is_vqa @classmethod def _A( cls , snake_case_ , snake_case_ , **snake_case_ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _A( self ): lowercase =copy.deepcopy(self.__dict__ ) lowercase =self.text_config.to_dict() lowercase =self.vision_config.to_dict() lowercase =self.__class__.model_type return output
72
0
"""simple docstring""" import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class UpperCamelCase__ : """simple docstring""" A__ : int = field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "Model type selected in the list: " + ", ".join(__SCREAMING_SNAKE_CASE )} ) A__ : List[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) A__ : Tuple = field( default=1_2_8 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) A__ : List[str] = field( default=1_2_8 , metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} , ) A__ : Any = field( default=6_4 , metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } , ) A__ : Dict = field( default=3_0 , metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } , ) A__ : Union[str, Any] = field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "Overwrite the cached training and evaluation sets"} ) A__ : List[str] = field( default=__SCREAMING_SNAKE_CASE , metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) A__ : List[str] = field( default=0.0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) A__ : Tuple = field( default=2_0 , metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) A__ : Dict = field( default=0 , metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } , ) A__ : Dict = field(default=1 , metadata={"help": "multiple threads for converting example to features"} ) class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" A__ : Union[str, Any] = "train" A__ : Optional[int] = "dev" class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" A__ : int = 4_2 A__ : int = 4_2 A__ : Any = 4_2 A__ : List[str] = 4_2 def __init__( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = Split.train , SCREAMING_SNAKE_CASE__ = False , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = "pt" , ) -> Any: A__ = args A__ = is_language_sensitive A__ = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(snake_case_ , snake_case_ ): try: A__ = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) A__ = mode # Load data features from cache or dataset file A__ = "v2" if args.version_2_with_negative else "v1" A__ = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}""" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. A__ = cached_features_file + ".lock" with FileLock(snake_case_ ): if os.path.exists(snake_case_ ) and not args.overwrite_cache: A__ = time.time() A__ = torch.load(snake_case_ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. A__ = self.old_features["features"] A__ = self.old_features.get("dataset" , snake_case_ ) A__ = self.old_features.get("examples" , snake_case_ ) logger.info( f"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"""Deleting cached file {cached_features_file} will allow dataset and examples to be cached in""" " future run" ) else: if mode == Split.dev: A__ = self.processor.get_dev_examples(args.data_dir ) else: A__ = self.processor.get_train_examples(args.data_dir ) A__ , A__ = squad_convert_examples_to_features( examples=self.examples , tokenizer=snake_case_ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=snake_case_ , ) A__ = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , snake_case_ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self ) -> str: return len(self.features ) def __getitem__( self , SCREAMING_SNAKE_CASE__ ) -> str: # Convert to Tensors and build dataset A__ = self.features[i] A__ = torch.tensor(feature.input_ids , dtype=torch.long ) A__ = torch.tensor(feature.attention_mask , dtype=torch.long ) A__ = torch.tensor(feature.token_type_ids , dtype=torch.long ) A__ = torch.tensor(feature.cls_index , dtype=torch.long ) A__ = torch.tensor(feature.p_mask , dtype=torch.float ) A__ = torch.tensor(feature.is_impossible , dtype=torch.float ) A__ = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: A__ = torch.tensor(feature.start_position , dtype=torch.long ) A__ = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
104
'''simple docstring''' def UpperCamelCase ( ) -> int: '''simple docstring''' return 1 def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else ten_pence(x - 1_0 ) + five_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else twenty_pence(x - 2_0 ) + ten_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else fifty_pence(x - 5_0 ) + twenty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else one_pound(x - 1_0_0 ) + fifty_pence(lowercase_ ) def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' return 0 if x < 0 else two_pound(x - 2_0_0 ) + one_pound(lowercase_ ) def UpperCamelCase ( lowercase_ : int = 2_0_0 ) -> int: '''simple docstring''' return two_pound(lowercase_ ) if __name__ == "__main__": print(solution(int(input().strip())))
72
0
"""simple docstring""" import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=16 , _lowerCamelCase=2 , _lowerCamelCase=0.02 , _lowerCamelCase=4 , ): """simple docstring""" UpperCAmelCase__ : List[str] = parent UpperCAmelCase__ : Dict = batch_size UpperCAmelCase__ : Tuple = seq_length UpperCAmelCase__ : Any = is_training UpperCAmelCase__ : Any = use_attention_mask UpperCAmelCase__ : Union[str, Any] = use_token_type_ids UpperCAmelCase__ : int = use_labels UpperCAmelCase__ : Union[str, Any] = vocab_size UpperCAmelCase__ : List[Any] = hidden_size UpperCAmelCase__ : Optional[Any] = num_hidden_layers UpperCAmelCase__ : str = num_attention_heads UpperCAmelCase__ : str = intermediate_size UpperCAmelCase__ : Dict = hidden_act UpperCAmelCase__ : int = hidden_dropout_prob UpperCAmelCase__ : List[Any] = attention_probs_dropout_prob UpperCAmelCase__ : List[Any] = max_position_embeddings UpperCAmelCase__ : Tuple = type_vocab_size UpperCAmelCase__ : Tuple = type_sequence_label_size UpperCAmelCase__ : List[Any] = initializer_range UpperCAmelCase__ : Union[str, Any] = num_choices def _a (self ): """simple docstring""" UpperCAmelCase__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ : Union[str, Any] = None if self.use_attention_mask: UpperCAmelCase__ : int = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : Optional[int] = None if self.use_token_type_ids: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ : List[str] = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case_ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _a (self ): """simple docstring""" UpperCAmelCase__ : Dict = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = config_and_inputs UpperCAmelCase__ : int = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class lowerCamelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def _a (self ): """simple docstring""" UpperCAmelCase__ : int = FlaxRoFormerModelTester(self ) @slow def _a (self ): """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase__ : List[str] = model_class_name.from_pretrained("""junnyu/roformer_chinese_small""" , from_pt=snake_case_ ) UpperCAmelCase__ : List[str] = model(np.ones((1, 1) ) ) self.assertIsNotNone(snake_case_ ) @require_flax class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = FlaxRoFormerForMaskedLM.from_pretrained("""junnyu/roformer_chinese_base""" ) UpperCAmelCase__ : Optional[int] = jnp.array([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase__ : Optional[Any] = model(snake_case_ )[0] UpperCAmelCase__ : List[Any] = 50000 UpperCAmelCase__ : Any = (1, 6, vocab_size) self.assertEqual(output.shape , snake_case_ ) UpperCAmelCase__ : Optional[Any] = jnp.array( [[[-0.1_205, -1.0_265, 0.2_922], [-1.5_134, 0.1_974, 0.1_519], [-5.0_135, -3.9_003, -0.8_404]]] ) self.assertTrue(jnp.allclose(output[:, :3, :3] , snake_case_ , atol=1e-4 ) )
182
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
0
import os import numpy import onnx def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> List[Any]: _lowercase : Union[str, Any] = a.name _lowercase : Tuple = b.name _lowercase : Dict = '' _lowercase : Union[str, Any] = '' _lowercase : Tuple = a == b _lowercase : Union[str, Any] = name_a _lowercase : Tuple = name_b return res def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(lowercase_ , lowercase_ ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , lowercase_ , lowercase_ ) _graph_replace_input_with(node_proto.attribute[1].g , lowercase_ , lowercase_ ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , lowercase_ , lowercase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Union[str, Any]: for n in graph_proto.node: _node_replace_input_with(lowercase_ , lowercase_ , lowercase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Tuple: _lowercase : Dict = list(model.graph.initializer ) _lowercase : List[str] = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i _lowercase : Dict = inits[i].name _lowercase : str = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , lowercase_ , lowercase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: _lowercase : Tuple = os.path.dirname(lowercase_ ) _lowercase : Any = os.path.basename(lowercase_ ) _lowercase : List[Any] = onnx.load(os.path.join(lowercase_ , lowercase_ ) ) _lowercase : Tuple = list(model.graph.initializer ) _lowercase : Tuple = set() _lowercase : int = {} _lowercase : List[Any] = [] _lowercase : int = 0 for i in range(len(lowercase_ ) ): if i in dup_set: continue for j in range(i + 1 , len(lowercase_ ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(lowercase_ ) dup_set.add(lowercase_ ) _lowercase : List[str] = inits[j].data_type _lowercase : List[Any] = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , lowercase_ ) total_reduced_size += mem_size _lowercase : Tuple = inits[i].name _lowercase : int = inits[j].name if name_i in dup_map: dup_map[name_i].append(lowercase_ ) else: _lowercase : Any = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) _lowercase : Dict = sorted(lowercase_ ) _remove_dup_initializers_from_model(lowercase_ , lowercase_ , lowercase_ ) _lowercase : Optional[int] = 'optimized_' + model_file_name _lowercase : Union[str, Any] = os.path.join(lowercase_ , lowercase_ ) onnx.save(lowercase_ , lowercase_ ) return new_model
89
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _UpperCAmelCase : Dict = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' _UpperCAmelCase : Union[str, Any] = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' _UpperCAmelCase : Dict = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=None , snake_case_=True , snake_case_=False ): if rouge_types is None: lowercase =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase =rouge_scorer.RougeScorer(rouge_types=snake_case_ , use_stemmer=snake_case_ ) if use_aggregator: lowercase =scoring.BootstrapAggregator() else: lowercase =[] for ref, pred in zip(snake_case_ , snake_case_ ): lowercase =scorer.score(snake_case_ , snake_case_ ) if use_aggregator: aggregator.add_scores(snake_case_ ) else: scores.append(snake_case_ ) if use_aggregator: lowercase =aggregator.aggregate() else: lowercase ={} for key in scores[0]: lowercase =[score[key] for score in scores] return result
72
0
from math import loga def __UpperCAmelCase( lowercase_ ): if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(lowercase_ , lowercase_ ): raise TypeError('''Input value must be a \'int\' type''' ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
114
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : str = '''▁''' _UpperCAmelCase : Union[str, Any] = {'''vocab_file''': '''spiece.model'''} _UpperCAmelCase : Union[str, Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _UpperCAmelCase : List[Any] = { '''google/pegasus-xsum''': 5_12, } _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = VOCAB_FILES_NAMES UpperCamelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ = ['input_ids', 'attention_mask'] def __init__( self , snake_case_ , snake_case_="<pad>" , snake_case_="</s>" , snake_case_="<unk>" , snake_case_="<mask_2>" , snake_case_="<mask_1>" , snake_case_=None , snake_case_=1_03 , snake_case_ = None , **snake_case_ , ): lowercase =offset if additional_special_tokens is not None: if not isinstance(snake_case_ , snake_case_ ): raise TypeError( f'additional_special_tokens should be of type {type(snake_case_ )}, but is' f' {type(snake_case_ )}' ) lowercase =( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(snake_case_ ) , self.offset - 1 ) ] if len(set(snake_case_ ) ) != len(snake_case_ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) lowercase =additional_special_tokens_extended else: lowercase =[mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2 , self.offset )] lowercase ={} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=snake_case_ , unk_token=snake_case_ , mask_token=snake_case_ , pad_token=snake_case_ , mask_token_sent=snake_case_ , offset=snake_case_ , additional_special_tokens=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) lowercase =mask_token_sent lowercase =vocab_file lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case_ ) # add special tokens to encoder dict lowercase ={ 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) lowercase ={v: k for k, v in self.encoder.items()} @property def _A( self ): return len(self.sp_model ) + self.offset def _A( self ): lowercase ={self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase =self.__dict__.copy() lowercase =None return state def __setstate__( self , snake_case_ ): lowercase =d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase ={} lowercase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _A( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def _A( self , snake_case_ ): if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] lowercase =self.sp_model.piece_to_id(snake_case_ ) return sp_id + self.offset def _A( self , snake_case_ ): if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: lowercase =self.sp_model.IdToPiece(index - self.offset ) return token def _A( self , snake_case_ ): lowercase =[] lowercase ='''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case_ ) + token lowercase =[] else: current_sub_tokens.append(snake_case_ ) out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def _A( self , snake_case_=False ): return 1 def _A( self , snake_case_ ): lowercase =set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _A( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return self._special_token_mask(snake_case_ ) elif token_ids_a is None: return self._special_token_mask(snake_case_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _A( self , snake_case_ , snake_case_=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _A( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowercase =os.path.join( snake_case_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , '''wb''' ) as fi: lowercase =self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
72
0
from __future__ import annotations def UpperCamelCase ( __lowercase : list[int | str] ): '''simple docstring''' create_state_space_tree(lowercase_ ,[] ,0 ,[0 for i in range(len(lowercase_ ) )] ) def UpperCamelCase ( __lowercase : list[int | str] ,__lowercase : list[int | str] ,__lowercase : int ,__lowercase : list[int] ,): '''simple docstring''' if index == len(lowercase_ ): print(lowercase_ ) return for i in range(len(lowercase_ ) ): if not index_used[i]: current_sequence.append(sequence[i] ) A_ : Optional[Any] = True create_state_space_tree(lowercase_ ,lowercase_ ,index + 1 ,lowercase_ ) current_sequence.pop() A_ : int = False _UpperCAmelCase = [3, 1, 2, 4] generate_all_permutations(sequence) _UpperCAmelCase = ["A", "B", "C"] generate_all_permutations(sequence_a)
558
'''simple docstring''' def UpperCamelCase ( lowercase_ : int , lowercase_ : int ) -> str: '''simple docstring''' return "\n".join( f'{number} * {i} = {number * i}' for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
72
0
'''simple docstring''' class a_ : def __init__( self , snake_case_ , snake_case_ ): _lowerCAmelCase : Dict = name _lowerCAmelCase : Union[str, Any] = val def __str__( self ): return f'{self.__class__.__name__}({self.name}, {self.val})' def __lt__( self , snake_case_ ): return self.val < other.val class a_ : def __init__( self , snake_case_ ): _lowerCAmelCase : Tuple = {} _lowerCAmelCase : Optional[int] = {} _lowerCAmelCase : Dict = self.build_heap(snake_case_ ) def __getitem__( self , snake_case_ ): return self.get_value(snake_case_ ) def __UpperCamelCase ( self , snake_case_ ): return (idx - 1) // 2 def __UpperCamelCase ( self , snake_case_ ): return idx * 2 + 1 def __UpperCamelCase ( self , snake_case_ ): return idx * 2 + 2 def __UpperCamelCase ( self , snake_case_ ): return self.heap_dict[key] def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : Optional[int] = len(snake_case_ ) - 1 _lowerCAmelCase : List[Any] = self.get_parent_idx(snake_case_ ) for idx, i in enumerate(snake_case_ ): _lowerCAmelCase : str = idx _lowerCAmelCase : Tuple = i.val for i in range(snake_case_ , -1 , -1 ): self.sift_down(snake_case_ , snake_case_ ) return array def __UpperCamelCase ( self , snake_case_ , snake_case_ ): while True: _lowerCAmelCase : str = self.get_left_child_idx(snake_case_ ) # noqa: E741 _lowerCAmelCase : Dict = self.get_right_child_idx(snake_case_ ) _lowerCAmelCase : Optional[int] = idx if l < len(snake_case_ ) and array[l] < array[idx]: _lowerCAmelCase : int = l if r < len(snake_case_ ) and array[r] < array[smallest]: _lowerCAmelCase : int = r if smallest != idx: _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = array[smallest], array[idx] ( ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ) : Union[str, Any] = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) _lowerCAmelCase : Optional[int] = smallest else: break def __UpperCamelCase ( self , snake_case_ ): _lowerCAmelCase : List[str] = self.get_parent_idx(snake_case_ ) while p >= 0 and self.heap[p] > self.heap[idx]: _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = self.heap[idx], self.heap[p] _lowerCAmelCase , _lowerCAmelCase : Dict = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) _lowerCAmelCase : Dict = p _lowerCAmelCase : List[str] = self.get_parent_idx(snake_case_ ) def __UpperCamelCase ( self ): return self.heap[0] def __UpperCamelCase ( self ): _lowerCAmelCase , _lowerCAmelCase : Optional[int] = self.heap[-1], self.heap[0] _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) _lowerCAmelCase : int = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def __UpperCamelCase ( self , snake_case_ ): self.heap.append(snake_case_ ) _lowerCAmelCase : Tuple = len(self.heap ) - 1 _lowerCAmelCase : Optional[Any] = node.val self.sift_up(len(self.heap ) - 1 ) def __UpperCamelCase ( self ): return len(self.heap ) == 0 def __UpperCamelCase ( self , snake_case_ , snake_case_ ): assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" _lowerCAmelCase : Dict = new_value _lowerCAmelCase : List[Any] = new_value self.sift_up(self.idx_of_element[node] ) UpperCamelCase_ = Node("""R""", -1) UpperCamelCase_ = Node("""B""", 6) UpperCamelCase_ = Node("""A""", 3) UpperCamelCase_ = Node("""X""", 1) UpperCamelCase_ = Node("""E""", 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array UpperCamelCase_ = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print("""Min Heap - before decrease key""") for i in my_min_heap.heap: print(i) print("""Min Heap - After decrease key of node [B -> -17]""") my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
384
'''simple docstring''' import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase ( lowercase_ : List[str] , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[Any]: '''simple docstring''' if isinstance(lowercase_ , lowercase_ ): lowercase =np.full((len(lowercase_ ), sequence_length, 2) , lowercase_ ) else: lowercase =np.full((len(lowercase_ ), sequence_length) , lowercase_ ) for i, tensor in enumerate(lowercase_ ): if padding_side == "right": if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] else: if isinstance(lowercase_ , lowercase_ ): lowercase =tensor[:sequence_length] else: lowercase =tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase ( lowercase_ : Optional[Any] ) -> str: '''simple docstring''' lowercase =ord(lowercase_ ) if (cp >= 3_3 and cp <= 4_7) or (cp >= 5_8 and cp <= 6_4) or (cp >= 9_1 and cp <= 9_6) or (cp >= 1_2_3 and cp <= 1_2_6): return True lowercase =unicodedata.category(lowercase_ ) if cat.startswith('''P''' ): return True return False @dataclass class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 42 UpperCamelCase__ = True UpperCamelCase__ = None UpperCamelCase__ = None UpperCamelCase__ = -1_00 UpperCamelCase__ = "pt" def _A( self , snake_case_ ): import torch lowercase ='''label''' if '''label''' in features[0].keys() else '''labels''' lowercase =[feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase =self.tokenizer.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' if labels is None else None , ) if labels is None: return batch lowercase =torch.tensor(batch['''entity_ids'''] ).shape[1] lowercase =self.tokenizer.padding_side if padding_side == "right": lowercase =[ list(snake_case_ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) for label in labels ] else: lowercase =[ [self.label_pad_token_id] * (sequence_length - len(snake_case_ )) + list(snake_case_ ) for label in labels ] lowercase =[feature['''ner_tags'''] for feature in features] lowercase =padding_tensor(snake_case_ , -1 , snake_case_ , snake_case_ ) lowercase =[feature['''original_entity_spans'''] for feature in features] lowercase =padding_tensor(snake_case_ , (-1, -1) , snake_case_ , snake_case_ ) lowercase ={k: torch.tensor(snake_case_ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
0
"""simple docstring""" import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore UpperCamelCase__ = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" UpperCamelCase__ = [file for file in filepaths if file != file.lower()] if upper_files: print(F"{len(upper_files)} files contain uppercase characters:") print('''\n'''.join(upper_files) + '''\n''') UpperCamelCase__ = [file for file in filepaths if ''' ''' in file] if space_files: print(F"{len(space_files)} files contain space characters:") print('''\n'''.join(space_files) + '''\n''') UpperCamelCase__ = [file for file in filepaths if '''-''' in file] if hyphen_files: print(F"{len(hyphen_files)} files contain hyphen characters:") print('''\n'''.join(hyphen_files) + '''\n''') UpperCamelCase__ = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"{len(nodir_files)} files are not in a directory:") print('''\n'''.join(nodir_files) + '''\n''') UpperCamelCase__ = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
227
'''simple docstring''' _UpperCAmelCase : Tuple = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def UpperCamelCase ( lowercase_ : bytes ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): lowercase =f'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(lowercase_ ) lowercase =''''''.join(bin(lowercase_ )[2:].zfill(8 ) for byte in data ) lowercase =len(lowercase_ ) % 6 != 0 if padding_needed: # The padding that will be added later lowercase =b'''=''' * ((6 - len(lowercase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowercase_ ) % 6) else: lowercase =b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowercase_ ) , 6 ) ).encode() + padding ) def UpperCamelCase ( lowercase_ : str ) -> bytes: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): lowercase =( '''argument should be a bytes-like object or ASCII string, ''' f'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(lowercase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowercase_ , lowercase_ ): try: lowercase =encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) lowercase =encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowercase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one lowercase =encoded_data[:-padding] lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: lowercase =''''''.join( bin(B64_CHARSET.index(lowercase_ ) )[2:].zfill(6 ) for char in encoded_data ) lowercase =[ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowercase_ ) , 8 ) ] return bytes(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
__magic_name__ : int = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' __magic_name__ : Dict = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __magic_name__ : Optional[int] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
280
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _UpperCAmelCase : Union[str, Any] = datasets.logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' _UpperCAmelCase : str = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' _UpperCAmelCase : Optional[int] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase ( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=False , lowercase_ : Optional[Any]=True , lowercase_ : Optional[Any]=False , lowercase_ : int="dummy_doc" ) -> str: '''simple docstring''' lowercase ={doc: key_lines} lowercase ={doc: sys_lines} lowercase ={} lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase =0 lowercase , lowercase =reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) lowercase , lowercase =reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: lowercase =reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters lowercase , lowercase =reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =reader.get_mention_assignments(lowercase_ , lowercase_ ) lowercase =(key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( '''Number of resulting singleton clusters in the key ''' f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' '''files, respectively''' ) return doc_coref_infos def UpperCamelCase ( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Tuple ) -> Dict: '''simple docstring''' lowercase =get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) lowercase ={} lowercase =0 lowercase =0 for name, metric in metrics: lowercase , lowercase , lowercase =evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , f'Recall: {recall * 1_0_0:.2f}' , f' Precision: {precision * 1_0_0:.2f}' , f' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: lowercase =(conll / 3) * 1_0_0 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({'''conll_score''': conll} ) return output_scores def UpperCamelCase ( lowercase_ : Any ) -> List[Any]: '''simple docstring''' lowercase =False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: lowercase =line.split()[5] if not parse_col == "-": lowercase =True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def _A( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): lowercase =[ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: lowercase =util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" lowercase =evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
72
0
'''simple docstring''' import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from datasets.features import ArrayaD, ClassLabel, Features, Image, Value from datasets.features.features import ArrayaDExtensionType, cast_to_python_objects from datasets.keyhash import DuplicatedKeysError, InvalidKeyError from .utils import require_pil class _lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def _a (self ): A_ : Any = pa.array(TypedSequence([1, 2, 3] ) ) self.assertEqual(arr.type , pa.intaa() ) def _a (self ): with self.assertRaises(snake_case_ ): A_ : Optional[int] = pa.array(TypedSequence([1, 2, 3] ) , type=pa.intaa() ) def _a (self ): with self.assertRaises(snake_case_ ): A_ : List[str] = pa.array(TypedSequence([1, 2, 3] , try_type=Value("""bool""" ) , type=Value("""int64""" ) ) ) def _a (self ): A_ : Union[str, Any] = pa.array(TypedSequence([1, 2, 3] , type=Value("""int32""" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _a (self ): with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): A_ : Dict = pa.array(TypedSequence(["""foo""", """bar"""] , type=Value("""int64""" ) ) ) def _a (self ): A_ : Any = pa.array(TypedSequence([1, 2, 3] , try_type=Value("""int32""" ) ) ) self.assertEqual(arr.type , pa.intaa() ) def _a (self ): A_ : str = pa.array(TypedSequence(["""foo""", """bar"""] , try_type=Value("""int64""" ) ) ) self.assertEqual(arr.type , pa.string() ) def _a (self ): A_ : List[str] = pa.array(TypedSequence([[[1, 2, 3]]] , type=ArrayaD((1, 3) , """int64""" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , """int64""" ) ) def _a (self ): with self.assertRaises((TypeError, pa.lib.ArrowInvalid) ): A_ : Optional[Any] = pa.array(TypedSequence(["""foo""", """bar"""] , type=ArrayaD((1, 3) , """int64""" ) ) ) def _a (self ): A_ : List[Any] = pa.array(TypedSequence([[[1, 2, 3]]] , try_type=ArrayaD((1, 3) , """int64""" ) ) ) self.assertEqual(arr.type , ArrayaDExtensionType((1, 3) , """int64""" ) ) def _a (self ): A_ : Dict = pa.array(TypedSequence(["""foo""", """bar"""] , try_type=ArrayaD((1, 3) , """int64""" ) ) ) self.assertEqual(arr.type , pa.string() ) @require_pil def _a (self ): import PIL.Image A_ : Any = PIL.Image.fromarray(np.arange(10 , dtype=np.uinta ).reshape(2 , 5 ) ) with patch( """datasets.arrow_writer.cast_to_python_objects""" , side_effect=snake_case_ ) as mock_cast_to_python_objects: A_ : Optional[int] = pa.array(TypedSequence([{"""path""": None, """bytes""": b"""image_bytes"""}, pil_image] , type=Image() ) ) A_, A_ : List[Any] = mock_cast_to_python_objects.call_args_list[-1] self.assertIn("""optimize_list_casting""" , snake_case_ ) self.assertFalse(kwargs["""optimize_list_casting"""] ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Union[str, Any] = pa.BufferReader(lowercase_ ) if isinstance(lowercase_ , pa.Buffer ) else pa.memory_map(lowercase_ ) A_ : str = pa.ipc.open_stream(lowercase_ ) A_ : Dict = f.read_all() assert len(pa_table.to_batches() ) == expected_num_chunks assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} del pa_table @pytest.mark.parametrize("""writer_batch_size""" , [None, 1, 10] ) @pytest.mark.parametrize( """fields""" , [None, {"""col_1""": pa.string(), """col_2""": pa.intaa()}, {"""col_1""": pa.string(), """col_2""": pa.intaa()}] ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Any = pa.BufferOutputStream() A_ : List[Any] = pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write({"""col_1""": """foo""", """col_2""": 1} ) writer.write({"""col_1""": """bar""", """col_2""": 2} ) A_, A_ : Dict = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: A_ : Tuple = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def a ( ): '''simple docstring''' A_ : Union[str, Any] = pa.BufferOutputStream() A_ : Tuple = Features({"""labels""": ClassLabel(names=["""neg""", """pos"""] )} ) with ArrowWriter(stream=lowercase_ , features=lowercase_ ) as writer: writer.write({"""labels""": 0} ) writer.write({"""labels""": 1} ) A_, A_ : Dict = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == features.arrow_schema assert writer._schema.metadata == features.arrow_schema.metadata A_ : int = pa.BufferReader(output.getvalue() ) A_ : Dict = pa.ipc.open_stream(lowercase_ ) A_ : Any = f.read_all() A_ : Union[str, Any] = pa_table.schema assert pa_table.num_rows == 2 assert schema == features.arrow_schema assert schema.metadata == features.arrow_schema.metadata assert features == Features.from_arrow_schema(lowercase_ ) @pytest.mark.parametrize("""writer_batch_size""" , [None, 1, 10] ) def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Tuple = pa.BufferOutputStream() with ArrowWriter( stream=lowercase_ , writer_batch_size=lowercase_ , hash_salt="""split_name""" , check_duplicates=lowercase_ , ) as writer: with pytest.raises(lowercase_ ): writer.write({"""col_1""": """foo""", """col_2""": 1} , key=[1, 2] ) A_, A_ : List[str] = writer.finalize() @pytest.mark.parametrize("""writer_batch_size""" , [None, 2, 10] ) def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Dict = pa.BufferOutputStream() with ArrowWriter( stream=lowercase_ , writer_batch_size=lowercase_ , hash_salt="""split_name""" , check_duplicates=lowercase_ , ) as writer: with pytest.raises(lowercase_ ): writer.write({"""col_1""": """foo""", """col_2""": 1} , key=10 ) writer.write({"""col_1""": """bar""", """col_2""": 2} , key=10 ) A_, A_ : List[Any] = writer.finalize() @pytest.mark.parametrize("""writer_batch_size""" , [None, 2, 10] ) def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Any = pa.BufferOutputStream() with ArrowWriter( stream=lowercase_ , writer_batch_size=lowercase_ , hash_salt="""split_name""" , check_duplicates=lowercase_ , ) as writer: writer.write({"""col_1""": """foo""", """col_2""": 1} , key=1 ) writer.write({"""col_1""": """bar""", """col_2""": 2} , key=2 ) A_, A_ : Union[str, Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("""writer_batch_size""" , [None, 1, 10] ) @pytest.mark.parametrize( """fields""" , [None, {"""col_1""": pa.string(), """col_2""": pa.intaa()}, {"""col_1""": pa.string(), """col_2""": pa.intaa()}] ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : str = pa.BufferOutputStream() A_ : Union[str, Any] = pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write_batch({"""col_1""": ["""foo""", """bar"""], """col_2""": [1, 2]} ) writer.write_batch({"""col_1""": [], """col_2""": []} ) A_, A_ : Union[str, Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: A_ : str = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("""writer_batch_size""" , [None, 1, 10] ) @pytest.mark.parametrize( """fields""" , [None, {"""col_1""": pa.string(), """col_2""": pa.intaa()}, {"""col_1""": pa.string(), """col_2""": pa.intaa()}] ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : List[Any] = pa.BufferOutputStream() A_ : Dict = pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write_table(pa.Table.from_pydict({"""col_1""": ["""foo""", """bar"""], """col_2""": [1, 2]} ) ) A_, A_ : Dict = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: A_ : List[str] = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) @pytest.mark.parametrize("""writer_batch_size""" , [None, 1, 10] ) @pytest.mark.parametrize( """fields""" , [None, {"""col_1""": pa.string(), """col_2""": pa.intaa()}, {"""col_1""": pa.string(), """col_2""": pa.intaa()}] ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Any = pa.BufferOutputStream() A_ : Union[str, Any] = pa.schema(lowercase_ ) if fields else None with ArrowWriter(stream=lowercase_ , schema=lowercase_ , writer_batch_size=lowercase_ ) as writer: writer.write_row(pa.Table.from_pydict({"""col_1""": ["""foo"""], """col_2""": [1]} ) ) writer.write_row(pa.Table.from_pydict({"""col_1""": ["""bar"""], """col_2""": [2]} ) ) A_, A_ : Any = writer.finalize() assert num_examples == 2 assert num_bytes > 0 if not fields: A_ : Tuple = {"""col_1""": pa.string(), """col_2""": pa.intaa()} assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(output.getvalue() , expected_num_chunks=num_examples if writer_batch_size == 1 else 1 ) def a ( ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: A_ : Optional[Any] = {"""col_1""": pa.string(), """col_2""": pa.intaa()} A_ : int = os.path.join(lowercase_ , """test.arrow""" ) with ArrowWriter(path=lowercase_ , schema=pa.schema(lowercase_ ) ) as writer: writer.write_batch({"""col_1""": ["""foo""", """bar"""], """col_2""": [1, 2]} ) A_, A_ : str = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert writer._schema == pa.schema(lowercase_ , metadata=writer._schema.metadata ) _check_output(lowercase_ , 1 ) def a ( lowerCamelCase__ ): '''simple docstring''' if pa.types.is_list(lowercase_ ): return get_base_dtype(arr_type.value_type ) else: return arr_type def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' if isinstance(lst[0] , lowercase_ ): change_first_primitive_element_in_list(lst[0] , lowercase_ ) else: A_ : Tuple = value @pytest.mark.parametrize("""optimized_int_type, expected_dtype""" , [(None, pa.intaa()), (Value("""int32""" ), pa.intaa())] ) @pytest.mark.parametrize("""sequence""" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : List[str] = pa.array(TypedSequence(lowercase_ , optimized_int_type=lowercase_ ) ) assert get_base_dtype(arr.type ) == expected_dtype @pytest.mark.parametrize( """col, expected_dtype""" , [ ("""attention_mask""", pa.inta()), ("""special_tokens_mask""", pa.inta()), ("""token_type_ids""", pa.inta()), ("""input_ids""", pa.intaa()), ("""other""", pa.intaa()), ] , ) @pytest.mark.parametrize("""sequence""" , [[1, 2, 3], [[1, 2, 3]], [[[1, 2, 3]]]] ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : str = pa.array(OptimizedTypedSequence(lowercase_ , col=lowercase_ ) ) assert get_base_dtype(arr.type ) == expected_dtype # not in range if col != "other": # avoids errors due to in-place modifications A_ : List[str] = copy.deepcopy(lowercase_ ) A_ : Dict = np.iinfo(expected_dtype.to_pandas_dtype() ).max + 1 change_first_primitive_element_in_list(lowercase_ , lowercase_ ) A_ : List[Any] = pa.array(OptimizedTypedSequence(lowercase_ , col=lowercase_ ) ) assert get_base_dtype(arr.type ) == pa.intaa() @pytest.mark.parametrize("""raise_exception""" , [False, True] ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : List[str] = str(tmp_path / """dataset-train.arrow""" ) try: with ArrowWriter(path=lowercase_ ) as writer: if raise_exception: raise pa.lib.ArrowInvalid() else: writer.stream.close() except pa.lib.ArrowInvalid: pass finally: assert writer.stream.closed def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Any = """mock://dataset-train.arrow""" with ArrowWriter(path=lowercase_ , storage_options=mockfs.storage_options ) as writer: assert isinstance(writer._fs , type(lowercase_ ) ) assert writer._fs.storage_options == mockfs.storage_options writer.write({"""col_1""": """foo""", """col_2""": 1} ) writer.write({"""col_1""": """bar""", """col_2""": 2} ) A_, A_ : Any = writer.finalize() assert num_examples == 2 assert num_bytes > 0 assert mockfs.exists(lowercase_ ) def a ( ): '''simple docstring''' A_ : Any = pa.BufferOutputStream() with ParquetWriter(stream=lowercase_ ) as writer: writer.write({"""col_1""": """foo""", """col_2""": 1} ) writer.write({"""col_1""": """bar""", """col_2""": 2} ) A_, A_ : Optional[Any] = writer.finalize() assert num_examples == 2 assert num_bytes > 0 A_ : Tuple = pa.BufferReader(output.getvalue() ) A_ : Tuple = pq.read_table(lowercase_ ) assert pa_table.to_pydict() == {"col_1": ["foo", "bar"], "col_2": [1, 2]} @require_pil @pytest.mark.parametrize("""embed_local_files""" , [False, True] ) def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' import PIL.Image A_ : List[str] = str(tmp_path / """test_image_rgb.jpg""" ) PIL.Image.fromarray(np.zeros((5, 5) , dtype=np.uinta ) ).save(lowercase_ , format="""png""" ) A_ : Optional[Any] = pa.BufferOutputStream() with ParquetWriter( stream=lowercase_ , features=Features({"""image""": Image()} ) , embed_local_files=lowercase_ ) as writer: writer.write({"""image""": image_path} ) writer.finalize() A_ : str = pa.BufferReader(output.getvalue() ) A_ : Tuple = pq.read_table(lowercase_ ) A_ : str = pa_table.to_pydict() if embed_local_files: assert isinstance(out["""image"""][0]["""path"""] , lowercase_ ) with open(lowercase_ , """rb""" ) as f: assert out["image"][0]["bytes"] == f.read() else: assert out["image"][0]["path"] == image_path assert out["image"][0]["bytes"] is None def a ( ): '''simple docstring''' A_ : List[str] = pa.schema([pa.field("""col_1""" , pa.string() , nullable=lowercase_ )] ) A_ : List[str] = pa.BufferOutputStream() with ArrowWriter(stream=lowercase_ ) as writer: writer._build_writer(inferred_schema=lowercase_ ) assert writer._schema == pa.schema([pa.field("""col_1""" , pa.string() )] )
667
'''simple docstring''' def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' if n == 1 or not isinstance(lowercase_ , lowercase_ ): return 0 elif n == 2: return 1 else: lowercase =[0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase ( lowercase_ : int ) -> int: '''simple docstring''' lowercase =0 lowercase =2 while digits < n: index += 1 lowercase =len(str(fibonacci(lowercase_ ) ) ) return index def UpperCamelCase ( lowercase_ : int = 1_0_0_0 ) -> int: '''simple docstring''' return fibonacci_digits_index(lowercase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
72
0
'''simple docstring''' from __future__ import annotations from math import pi def a_ ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if inductance < 0: raise ValueError("Inductance cannot be negative" ) if frequency < 0: raise ValueError("Frequency cannot be negative" ) if reactance < 0: raise ValueError("Inductive reactance cannot be negative" ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
452
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : str = logging.get_logger(__name__) _UpperCAmelCase : Any = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'marian' UpperCamelCase__ = ['past_key_values'] UpperCamelCase__ = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , snake_case_=5_81_01 , snake_case_=None , snake_case_=10_24 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=12 , snake_case_=40_96 , snake_case_=16 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=True , snake_case_=True , snake_case_="gelu" , snake_case_=10_24 , snake_case_=0.1 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=5_81_00 , snake_case_=False , snake_case_=5_81_00 , snake_case_=0 , snake_case_=0 , snake_case_=True , **snake_case_ , ): lowercase =vocab_size lowercase =decoder_vocab_size or vocab_size lowercase =max_position_embeddings lowercase =d_model lowercase =encoder_ffn_dim lowercase =encoder_layers lowercase =encoder_attention_heads lowercase =decoder_ffn_dim lowercase =decoder_layers lowercase =decoder_attention_heads lowercase =dropout lowercase =attention_dropout lowercase =activation_dropout lowercase =activation_function lowercase =init_std lowercase =encoder_layerdrop lowercase =decoder_layerdrop lowercase =use_cache lowercase =encoder_layers lowercase =scale_embedding # scale factor will be sqrt(d_model) if True lowercase =share_encoder_decoder_embeddings super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , is_encoder_decoder=snake_case_ , decoder_start_token_id=snake_case_ , forced_eos_token_id=snake_case_ , **snake_case_ , ) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase ={0: '''batch'''} lowercase ={0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowercase ={0: '''batch''', 1: '''decoder_sequence'''} lowercase ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} else: lowercase =OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def _A( self ): if self.task in ["default", "seq2seq-lm"]: lowercase =super().outputs else: lowercase =super(snake_case_ , self ).outputs if self.use_past: lowercase , lowercase =self.num_layers for i in range(snake_case_ ): lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} lowercase ={0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # Generate decoder inputs lowercase =seq_length if not self.use_past else 1 lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) lowercase ={f'decoder_{name}': tensor for name, tensor in decoder_inputs.items()} lowercase =dict(**snake_case_ , **snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape lowercase =common_inputs['''decoder_input_ids'''].shape[1] lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =decoder_seq_length + 3 lowercase =( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) lowercase =torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(snake_case_ , snake_case_ )] , dim=1 ) lowercase =[] # If the number of encoder and decoder layers are present in the model configuration, both are considered lowercase , lowercase =self.num_layers lowercase =min(snake_case_ , snake_case_ ) lowercase =max(snake_case_ , snake_case_ ) - min_num_layers lowercase ='''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(snake_case_ ): common_inputs["past_key_values"].append( ( torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), torch.zeros(snake_case_ ), ) ) # TODO: test this. lowercase =encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(snake_case_ , snake_case_ ): common_inputs["past_key_values"].append((torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): lowercase =self._generate_dummy_inputs_for_encoder_and_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase , lowercase =common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase =seqlen + 2 lowercase , lowercase =self.num_layers lowercase , lowercase =self.num_attention_heads lowercase =( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) lowercase =common_inputs['''attention_mask'''].dtype lowercase =torch.cat( [common_inputs['''attention_mask'''], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) lowercase =[ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(snake_case_ ) ] return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase =tokenizer.num_special_tokens_to_add(snake_case_ ) lowercase =compute_effective_axis_dimension( snake_case_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=snake_case_ ) # Generate dummy inputs according to compute batch and sequence lowercase =[''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size lowercase =dict(tokenizer(snake_case_ , return_tensors=snake_case_ ) ) return common_inputs def _A( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): if self.task in ["default", "seq2seq-lm"]: lowercase =self._generate_dummy_inputs_for_default_and_seqaseq_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) else: lowercase =self._generate_dummy_inputs_for_causal_lm( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) return common_inputs def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): if self.task in ["default", "seq2seq-lm"]: lowercase =super()._flatten_past_key_values_(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) else: lowercase =super(snake_case_ , self )._flatten_past_key_values_( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) @property def _A( self ): return 1E-4
72
0
"""simple docstring""" import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ = None, ) -> int: a_ : Dict = {} if train_file is not None: a_ : List[str] = [train_file] if eval_file is not None: a_ : Union[str, Any] = [eval_file] if test_file is not None: a_ : List[str] = [test_file] a_ : Tuple = datasets.load_dataset("csv", data_files=lowercase_ ) a_ : str = list(ds[list(files.keys() )[0]].features.keys() ) a_ : Optional[Any] = features_name.pop(lowercase_ ) a_ : Optional[Any] = list(set(ds[list(files.keys() )[0]][label_name] ) ) a_ : Tuple = {label: i for i, label in enumerate(lowercase_ )} a_ : str = tokenizer.model_input_names a_ : Dict = {} if len(lowercase_ ) == 1: for k in files.keys(): a_ : Optional[int] = ds[k].map( lambda SCREAMING_SNAKE_CASE__ : tokenizer.batch_encode_plus( example[features_name[0]], truncation=lowercase_, max_length=lowercase_, padding="max_length" ), batched=lowercase_, ) elif len(lowercase_ ) == 2: for k in files.keys(): a_ : Tuple = ds[k].map( lambda SCREAMING_SNAKE_CASE__ : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]), truncation=lowercase_, max_length=lowercase_, padding="max_length", ), batched=lowercase_, ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: a_ : Union[str, Any] = {k: v for k, v in ex.items() if k in input_names} a_ : Union[str, Any] = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: a_ : Optional[Any] = {k: v for k, v in ex.items() if k in input_names} a_ : Optional[Any] = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: a_ : Any = {k: v for k, v in ex.items() if k in input_names} a_ : Optional[int] = labelaid[ex[label_name]] yield (d, label) a_ : List[Any] = ( tf.data.Dataset.from_generator( lowercase_, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: a_ : List[Any] = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) a_ : Union[str, Any] = ( tf.data.Dataset.from_generator( lowercase_, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: a_ : Dict = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) a_ : Optional[Any] = ( tf.data.Dataset.from_generator( lowercase_, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: a_ : Optional[Any] = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid SCREAMING_SNAKE_CASE_ = logging.getLogger(__name__) @dataclass class snake_case_ : __lowerCAmelCase = field(metadata={"help": "Which column contains the label"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "The path of the training file"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "The path of the development file"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "The path of the test file"} ) __lowerCAmelCase = field( default=1_2_8 ,metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } ,) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Overwrite the cached training and evaluation sets"} ) @dataclass class snake_case_ : __lowerCAmelCase = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Pretrained config name or path if not the same as model_name"} ) __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) __lowerCAmelCase = field(default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Set this flag to use fast tokenization."} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. __lowerCAmelCase = field( default=__SCREAMING_SNAKE_CASE ,metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} ,) def lowerCAmelCase_ ( ) -> Optional[int]: a_ : Dict = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) a_ , a_ , a_ : Optional[int] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" " --overwrite_output_dir to overcome." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO, ) logger.info( F"""n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, """ F"""16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. a_ : str = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path, cache_dir=model_args.cache_dir, ) a_ , a_ , a_ , a_ : Union[str, Any] = get_tfds( train_file=data_args.train_file, eval_file=data_args.dev_file, test_file=data_args.test_file, tokenizer=lowercase_, label_column_id=data_args.label_column_id, max_seq_length=data_args.max_seq_length, ) a_ : int = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path, num_labels=len(lowercase_ ), labelaid=lowercase_, idalabel={id: label for label, id in labelaid.items()}, finetuning_task="text-classification", cache_dir=model_args.cache_dir, ) with training_args.strategy.scope(): a_ : str = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path, from_pt=bool(".bin" in model_args.model_name_or_path ), config=lowercase_, cache_dir=model_args.cache_dir, ) def compute_metrics(SCREAMING_SNAKE_CASE__ ) -> Dict: a_ : Any = np.argmax(p.predictions, axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer a_ : Optional[int] = TFTrainer( model=lowercase_, args=lowercase_, train_dataset=lowercase_, eval_dataset=lowercase_, compute_metrics=lowercase_, ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation a_ : Dict = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) a_ : Optional[Any] = trainer.evaluate() a_ : int = os.path.join(training_args.output_dir, "eval_results.txt" ) with open(lowercase_, "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(F""" {key} = {value}""" ) writer.write(F"""{key} = {value}\n""" ) results.update(lowercase_ ) return results if __name__ == "__main__": main()
237
'''simple docstring''' import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _UpperCAmelCase : Dict = parse(importlib.metadata.version('''torch''')) def UpperCamelCase ( lowercase_ : Union[str, Version] , lowercase_ : str , lowercase_ : str ) -> List[Any]: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f'`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}' ) lowercase =STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): lowercase =parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def UpperCamelCase ( lowercase_ : str , lowercase_ : str ) -> Union[str, Any]: '''simple docstring''' return compare_versions(lowercase_ , lowercase_ , lowercase_ )
72
0
"""simple docstring""" from math import pi, sqrt, tan def _lowerCamelCase ( UpperCAmelCase_ : float ) -> float: """simple docstring""" if side_length < 0: raise ValueError("surface_area_cube() only accepts non-negative values" ) return 6 * side_length**2 def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if length < 0 or breadth < 0 or height < 0: raise ValueError("surface_area_cuboid() only accepts non-negative values" ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def _lowerCamelCase ( UpperCAmelCase_ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError("surface_area_sphere() only accepts non-negative values" ) return 4 * pi * radius**2 def _lowerCamelCase ( UpperCAmelCase_ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError("surface_area_hemisphere() only accepts non-negative values" ) return 3 * pi * radius**2 def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if radius < 0 or height < 0: raise ValueError("surface_area_cone() only accepts non-negative values" ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( "surface_area_conical_frustum() only accepts non-negative values" ) A__ = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if radius < 0 or height < 0: raise ValueError("surface_area_cylinder() only accepts non-negative values" ) return 2 * pi * radius * (height + radius) def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if torus_radius < 0 or tube_radius < 0: raise ValueError("surface_area_torus() only accepts non-negative values" ) if torus_radius < tube_radius: raise ValueError( "surface_area_torus() does not support spindle or self intersecting tori" ) return 4 * pow(lowercase_, 2 ) * torus_radius * tube_radius def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if length < 0 or width < 0: raise ValueError("area_rectangle() only accepts non-negative values" ) return length * width def _lowerCamelCase ( UpperCAmelCase_ : float ) -> float: """simple docstring""" if side_length < 0: raise ValueError("area_square() only accepts non-negative values" ) return side_length**2 def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if base < 0 or height < 0: raise ValueError("area_triangle() only accepts non-negative values" ) return (base * height) / 2 def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError("area_triangle_three_sides() only accepts non-negative values" ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError("Given three sides do not form a triangle" ) A__ = (sidea + sidea + sidea) / 2 A__ = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if base < 0 or height < 0: raise ValueError("area_parallelogram() only accepts non-negative values" ) return base * height def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if basea < 0 or basea < 0 or height < 0: raise ValueError("area_trapezium() only accepts non-negative values" ) return 1 / 2 * (basea + basea) * height def _lowerCamelCase ( UpperCAmelCase_ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError("area_circle() only accepts non-negative values" ) return pi * radius**2 def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if radius_x < 0 or radius_y < 0: raise ValueError("area_ellipse() only accepts non-negative values" ) return pi * radius_x * radius_y def _lowerCamelCase ( UpperCAmelCase_ : float, UpperCAmelCase_ : float ) -> float: """simple docstring""" if diagonal_a < 0 or diagonal_a < 0: raise ValueError("area_rhombus() only accepts non-negative values" ) return 1 / 2 * diagonal_a * diagonal_a def _lowerCamelCase ( UpperCAmelCase_ : int, UpperCAmelCase_ : float ) -> float: """simple docstring""" if not isinstance(lowercase_, lowercase_ ) or sides < 3: raise ValueError( "area_reg_polygon() only accepts integers greater than or \\nequal to three as number of sides" ) elif length < 0: raise ValueError( "area_reg_polygon() only accepts non-negative values as \\nlength of a side" ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print("""[DEMO] Areas of various geometric shapes: \n""") print(f'Rectangle: {area_rectangle(10, 20) = }') print(f'Square: {area_square(10) = }') print(f'Triangle: {area_triangle(10, 10) = }') print(f'Triangle: {area_triangle_three_sides(5, 12, 13) = }') print(f'Parallelogram: {area_parallelogram(10, 20) = }') print(f'Rhombus: {area_rhombus(10, 20) = }') print(f'Trapezium: {area_trapezium(10, 20, 30) = }') print(f'Circle: {area_circle(20) = }') print(f'Ellipse: {area_ellipse(10, 20) = }') print("""\nSurface Areas of various geometric shapes: \n""") print(f'Cube: {surface_area_cube(20) = }') print(f'Cuboid: {surface_area_cuboid(10, 20, 30) = }') print(f'Sphere: {surface_area_sphere(20) = }') print(f'Hemisphere: {surface_area_hemisphere(20) = }') print(f'Cone: {surface_area_cone(10, 20) = }') print(f'Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }') print(f'Cylinder: {surface_area_cylinder(10, 20) = }') print(f'Torus: {surface_area_torus(20, 10) = }') print(f'Equilateral Triangle: {area_reg_polygon(3, 10) = }') print(f'Square: {area_reg_polygon(4, 10) = }') print(f'Reqular Pentagon: {area_reg_polygon(5, 10) = }')
104
'''simple docstring''' from __future__ import annotations import time import numpy as np _UpperCAmelCase : int = [8, 5, 9, 7] _UpperCAmelCase : List[str] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _UpperCAmelCase : Union[str, Any] = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class __magic_name__ : def __init__( self , snake_case_ , snake_case_ , snake_case_ , ): lowercase =claim_vector lowercase =allocated_resources_table lowercase =maximum_claim_table def _A( self ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _A( self ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _A( self ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(snake_case_ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _A( self ): return {self.__need().index(snake_case_ ): i for i in self.__need()} def _A( self , **snake_case_ ): lowercase =self.__need() lowercase =self.__allocated_resources_table lowercase =self.__available_resources() lowercase =self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('''_''' * 50 + '''\n''' ) while need_list: lowercase =False for each_need in need_list: lowercase =True for index, need in enumerate(snake_case_ ): if need > available_resources[index]: lowercase =False break if execution: lowercase =True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: lowercase =original_need_index print(f'Process {process_number + 1} is executing.' ) # remove the process run from stack need_list.remove(snake_case_ ) # update available/freed resources stack lowercase =np.array(snake_case_ ) + np.array( alloc_resources_table[process_number] ) print( '''Updated available resource stack for processes: ''' + ''' '''.join([str(snake_case_ ) for x in available_resources] ) ) break if safe: print('''The process is in a safe state.\n''' ) else: print('''System in unsafe state. Aborting...\n''' ) break def _A( self ): print(''' ''' * 9 + '''Allocated Resource Table''' ) for item in self.__allocated_resources_table: print( f'P{self.__allocated_resources_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print(''' ''' * 9 + '''System Resource Table''' ) for item in self.__maximum_claim_table: print( f'P{self.__maximum_claim_table.index(snake_case_ ) + 1}' + ''' '''.join(f'{it:>8}' for it in item ) + '''\n''' ) print( '''Current Usage by Active Processes: ''' + ''' '''.join(str(snake_case_ ) for x in self.__claim_vector ) ) print( '''Initial Available Resources: ''' + ''' '''.join(str(snake_case_ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase : '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=12 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=32 , _lowerCamelCase=32 , _lowerCamelCase=2 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=0.02 , _lowerCamelCase=0 , _lowerCamelCase=None , ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Dict = batch_size UpperCAmelCase__ : List[str] = seq_length UpperCAmelCase__ : Any = is_training UpperCAmelCase__ : Any = use_input_mask UpperCAmelCase__ : List[str] = use_labels UpperCAmelCase__ : Union[str, Any] = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : Any = projection_dim UpperCAmelCase__ : Dict = num_hidden_layers UpperCAmelCase__ : List[str] = num_attention_heads UpperCAmelCase__ : List[str] = intermediate_size UpperCAmelCase__ : Optional[Any] = dropout UpperCAmelCase__ : Optional[Any] = attention_dropout UpperCAmelCase__ : Union[str, Any] = max_position_embeddings UpperCAmelCase__ : Any = initializer_range UpperCAmelCase__ : List[str] = scope UpperCAmelCase__ : List[Any] = bos_token_id def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ : Union[str, Any] = None if self.use_input_mask: UpperCAmelCase__ : Dict = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: UpperCAmelCase__ : int = input_mask.numpy() UpperCAmelCase__ , UpperCAmelCase__ : List[str] = input_mask.shape UpperCAmelCase__ : List[Any] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(snake_case_ ): UpperCAmelCase__ : Optional[int] = 1 UpperCAmelCase__ : List[Any] = 0 UpperCAmelCase__ : int = self.get_config() return config, input_ids, tf.convert_to_tensor(snake_case_ ) def _a (self ): """simple docstring""" return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Dict = TFBlipTextModel(config=snake_case_ ) UpperCAmelCase__ : Tuple = model(snake_case_ , attention_mask=snake_case_ , training=snake_case_ ) UpperCAmelCase__ : List[Any] = model(snake_case_ , training=snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _a (self ): """simple docstring""" UpperCAmelCase__ : int = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = config_and_inputs UpperCAmelCase__ : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class lowerCamelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = (TFBlipTextModel,) if is_tf_available() else () SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _a (self ): """simple docstring""" UpperCAmelCase__ : str = BlipTextModelTester(self ) UpperCAmelCase__ : List[Any] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _a (self ): """simple docstring""" self.config_tester.run_common_tests() def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def _a (self ): """simple docstring""" pass def _a (self ): """simple docstring""" pass @unittest.skip(reason="""Blip does not use inputs_embeds""" ) def _a (self ): """simple docstring""" pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def _a (self ): """simple docstring""" pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def _a (self ): """simple docstring""" pass @slow def _a (self ): """simple docstring""" for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : Dict = TFBlipTextModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _a (self , _lowerCamelCase=True ): """simple docstring""" super().test_pt_tf_model_equivalence(allow_missing_keys=snake_case_ )
182
'''simple docstring''' import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version _UpperCAmelCase : Dict = version.parse(importlib_metadata.version('''nltk''')) if NLTK_VERSION >= version.Version('''3.6.4'''): from nltk import word_tokenize _UpperCAmelCase : Dict = '''\ @inproceedings{banarjee2005, title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments}, author = {Banerjee, Satanjeev and Lavie, Alon}, booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization}, month = jun, year = {2005}, address = {Ann Arbor, Michigan}, publisher = {Association for Computational Linguistics}, url = {https://www.aclweb.org/anthology/W05-0909}, pages = {65--72}, } ''' _UpperCAmelCase : Union[str, Any] = '''\ METEOR, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machine-produced translation and human-produced reference translations. Unigrams can be matched based on their surface forms, stemmed forms, and meanings; furthermore, METEOR can be easily extended to include more advanced matching strategies. Once all generalized unigram matches between the two strings have been found, METEOR computes a score for this matching using a combination of unigram-precision, unigram-recall, and a measure of fragmentation that is designed to directly capture how well-ordered the matched words in the machine translation are in relation to the reference. METEOR gets an R correlation value of 0.347 with human evaluation on the Arabic data and 0.331 on the Chinese data. This is shown to be an improvement on using simply unigram-precision, unigram-recall and their harmonic F1 combination. ''' _UpperCAmelCase : Tuple = ''' Computes METEOR score of translated segments against one or more references. Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. alpha: Parameter for controlling relative weights of precision and recall. default: 0.9 beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3 gamma: Relative weight assigned to fragmentation penalty. default: 0.5 Returns: \'meteor\': meteor score. Examples: >>> meteor = datasets.load_metric(\'meteor\') >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"] >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"] >>> results = meteor.compute(predictions=predictions, references=references) >>> print(round(results["meteor"], 4)) 0.6944 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def _A( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] , reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] , ) def _A( self , snake_case_ ): import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def _A( self , snake_case_ , snake_case_ , snake_case_=0.9 , snake_case_=3 , snake_case_=0.5 ): if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase =[ meteor_score.single_meteor_score( word_tokenize(snake_case_ ) , word_tokenize(snake_case_ ) , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] else: lowercase =[ meteor_score.single_meteor_score(snake_case_ , snake_case_ , alpha=snake_case_ , beta=snake_case_ , gamma=snake_case_ ) for ref, pred in zip(snake_case_ , snake_case_ ) ] return {"meteor": np.mean(snake_case_ )}
72
0
def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> float: if principal <= 0: raise Exception('Principal borrowed must be > 0' ) if rate_per_annum < 0: raise Exception('Rate of interest must be >= 0' ) if years_to_repay <= 0 or not isinstance(lowercase_ , lowercase_ ): raise Exception('Years to repay must be an integer > 0' ) # Yearly rate is divided by 12 to get monthly rate _lowercase : List[Any] = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly _lowercase : Dict = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
89
'''simple docstring''' import sys _UpperCAmelCase : Dict = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCamelCase ( lowercase_ : str = N ) -> int: '''simple docstring''' lowercase =-sys.maxsize - 1 for i in range(len(lowercase_ ) - 1_2 ): lowercase =1 for j in range(1_3 ): product *= int(n[i + j] ) if product > largest_product: lowercase =product return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
72
0
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _lowerCamelCase = parse(importlib.metadata.version('torch')) def __UpperCAmelCase( lowercase_ , lowercase_ , lowercase_ ): if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) _lowerCamelCase : Optional[int] = STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase_ , lowercase_ ): _lowerCamelCase : Union[str, Any] = parse(importlib.metadata.version(lowercase_ ) ) return operation(lowercase_ , parse(lowercase_ ) ) def __UpperCAmelCase( lowercase_ , lowercase_ ): return compare_versions(lowercase_ , lowercase_ , lowercase_ )
114
'''simple docstring''' import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _UpperCAmelCase : Any = logging.get_logger(__name__) class __magic_name__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , *snake_case_ , **snake_case_ ): warnings.warn( '''The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use YolosImageProcessor instead.''' , snake_case_ , ) super().__init__(*snake_case_ , **snake_case_ )
72
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { '''google/canine-s''': '''https://huggingface.co/google/canine-s/resolve/main/config.json''', # See all CANINE models at https://huggingface.co/models?filter=canine } class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCamelCase_ = '''canine''' def __init__( self , lowercase=7_6_8 , lowercase=1_2 , lowercase=1_2 , lowercase=3_0_7_2 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=1_6_3_8_4 , lowercase=1_6 , lowercase=0.02 , lowercase=1E-12 , lowercase=0 , lowercase=0Xe_000 , lowercase=0Xe_001 , lowercase=4 , lowercase=4 , lowercase=8 , lowercase=1_6_3_8_4 , lowercase=1_2_8 , **lowercase , ): """simple docstring""" super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) A_ : List[str] = max_position_embeddings A_ : Dict = hidden_size A_ : Dict = num_hidden_layers A_ : Dict = num_attention_heads A_ : List[Any] = intermediate_size A_ : int = hidden_act A_ : Tuple = hidden_dropout_prob A_ : Any = attention_probs_dropout_prob A_ : Any = initializer_range A_ : Dict = type_vocab_size A_ : Dict = layer_norm_eps # Character config: A_ : Tuple = downsampling_rate A_ : int = upsampling_kernel_size A_ : Union[str, Any] = num_hash_functions A_ : Optional[Any] = num_hash_buckets A_ : Tuple = local_transformer_stride
558
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = 'encodec' def __init__( self , snake_case_=[1.5, 3.0, 6.0, 12.0, 24.0] , snake_case_=2_40_00 , snake_case_=1 , snake_case_=False , snake_case_=None , snake_case_=None , snake_case_=1_28 , snake_case_=32 , snake_case_=1 , snake_case_=[8, 5, 4, 2] , snake_case_="weight_norm" , snake_case_=7 , snake_case_=7 , snake_case_=3 , snake_case_=2 , snake_case_=True , snake_case_="reflect" , snake_case_=2 , snake_case_=2 , snake_case_=1.0 , snake_case_=10_24 , snake_case_=None , snake_case_=True , **snake_case_ , ): lowercase =target_bandwidths lowercase =sampling_rate lowercase =audio_channels lowercase =normalize lowercase =chunk_length_s lowercase =overlap lowercase =hidden_size lowercase =num_filters lowercase =num_residual_layers lowercase =upsampling_ratios lowercase =norm_type lowercase =kernel_size lowercase =last_kernel_size lowercase =residual_kernel_size lowercase =dilation_growth_rate lowercase =use_causal_conv lowercase =pad_mode lowercase =compress lowercase =num_lstm_layers lowercase =trim_right_ratio lowercase =codebook_size lowercase =codebook_dim if codebook_dim is not None else hidden_size lowercase =use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def _A( self ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A( self ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A( self ): lowercase =np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A( self ): return int(10_00 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
72
0
'''simple docstring''' import argparse from collections import defaultdict import yaml UpperCamelCase_ = '''docs/source/en/_toctree.yml''' def _UpperCAmelCase ( _lowerCamelCase : int ) -> Dict: _lowerCAmelCase : Any = defaultdict(lowercase_ ) _lowerCAmelCase : int = [] _lowerCAmelCase : List[Any] = [] for doc in doc_list: if "local" in doc: counts[doc["local"]] += 1 if doc["title"].lower() == "overview": overview_doc.append({"""local""": doc["""local"""], """title""": doc["""title"""]} ) else: new_doc_list.append(lowercase_ ) _lowerCAmelCase : Optional[int] = new_doc_list _lowerCAmelCase : Tuple = [key for key, value in counts.items() if value > 1] _lowerCAmelCase : List[str] = [] for duplicate_key in duplicates: _lowerCAmelCase : Union[str, Any] = list({doc["""title"""] for doc in doc_list if doc["""local"""] == duplicate_key} ) if len(lowercase_ ) > 1: raise ValueError( f'{duplicate_key} is present several times in the documentation table of content at ' """`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the """ """others.""" ) # Only add this once new_doc.append({"""local""": duplicate_key, """title""": titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in doc_list if """local""" not in counts or counts[doc["""local"""]] == 1] ) _lowerCAmelCase : Dict = sorted(lowercase_ , key=lambda _lowerCamelCase : s["title"].lower() ) # "overview" gets special treatment and is always first if len(lowercase_ ) > 1: raise ValueError("""{doc_list} has two \'overview\' docs which is not allowed.""" ) overview_doc.extend(lowercase_ ) # Sort return overview_doc def _UpperCAmelCase ( _lowerCamelCase : Union[str, Any]=False ) -> List[str]: with open(lowercase_ , encoding="""utf-8""" ) as f: _lowerCAmelCase : Any = yaml.safe_load(f.read() ) # Get to the API doc _lowerCAmelCase : Dict = 0 while content[api_idx]["title"] != "API": api_idx += 1 _lowerCAmelCase : Tuple = content[api_idx]["""sections"""] # Then to the model doc _lowerCAmelCase : List[str] = 0 while api_doc[scheduler_idx]["title"] != "Schedulers": scheduler_idx += 1 _lowerCAmelCase : Dict = api_doc[scheduler_idx]["""sections"""] _lowerCAmelCase : Any = clean_doc_toc(lowercase_ ) _lowerCAmelCase : Union[str, Any] = False if new_scheduler_doc != scheduler_doc: _lowerCAmelCase : Optional[Any] = True if overwrite: _lowerCAmelCase : str = new_scheduler_doc if diff: if overwrite: _lowerCAmelCase : int = api_doc with open(lowercase_ , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(lowercase_ , allow_unicode=lowercase_ ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) def _UpperCAmelCase ( _lowerCamelCase : str=False ) -> Any: with open(lowercase_ , encoding="""utf-8""" ) as f: _lowerCAmelCase : int = yaml.safe_load(f.read() ) # Get to the API doc _lowerCAmelCase : Dict = 0 while content[api_idx]["title"] != "API": api_idx += 1 _lowerCAmelCase : Tuple = content[api_idx]["""sections"""] # Then to the model doc _lowerCAmelCase : Tuple = 0 while api_doc[pipeline_idx]["title"] != "Pipelines": pipeline_idx += 1 _lowerCAmelCase : Dict = False _lowerCAmelCase : Dict = api_doc[pipeline_idx]["""sections"""] _lowerCAmelCase : Tuple = [] # sort sub pipeline docs for pipeline_doc in pipeline_docs: if "section" in pipeline_doc: _lowerCAmelCase : Union[str, Any] = pipeline_doc["""section"""] _lowerCAmelCase : Optional[int] = clean_doc_toc(lowercase_ ) if overwrite: _lowerCAmelCase : int = new_sub_pipeline_doc new_pipeline_docs.append(lowercase_ ) # sort overall pipeline doc _lowerCAmelCase : Optional[int] = clean_doc_toc(lowercase_ ) if new_pipeline_docs != pipeline_docs: _lowerCAmelCase : Union[str, Any] = True if overwrite: _lowerCAmelCase : Dict = new_pipeline_docs if diff: if overwrite: _lowerCAmelCase : int = api_doc with open(lowercase_ , """w""" , encoding="""utf-8""" ) as f: f.write(yaml.dump(lowercase_ , allow_unicode=lowercase_ ) ) else: raise ValueError( """The model doc part of the table of content is not properly sorted, run `make style` to fix this.""" ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") UpperCamelCase_ = parser.parse_args() check_scheduler_doc(args.fix_and_overwrite) check_pipeline_doc(args.fix_and_overwrite)
384
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : int = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[Any] = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Dict = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
72
0
"""simple docstring""" import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( '''The `image_to_image.py` script is outdated. Please use directly `from diffusers import''' ''' StableDiffusionImg2ImgPipeline` instead.''' )
227
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: raise ValueError('''Capacitance cannot be 0 or negative''' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
72
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ : Any = logging.get_logger(__name__) __magic_name__ : Dict = { '''google/pix2struct-textcaps-base''': ( '''https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json''' ), } class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case__ = """pix2struct_text_model""" snake_case__ = ["""past_key_values"""] snake_case__ = { """hidden_size""": """hidden_size""", """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self : int , _SCREAMING_SNAKE_CASE : List[str]=5_0244 , _SCREAMING_SNAKE_CASE : Optional[Any]=768 , _SCREAMING_SNAKE_CASE : List[Any]=64 , _SCREAMING_SNAKE_CASE : Tuple=2048 , _SCREAMING_SNAKE_CASE : Dict=12 , _SCREAMING_SNAKE_CASE : Dict=12 , _SCREAMING_SNAKE_CASE : int=32 , _SCREAMING_SNAKE_CASE : Any=128 , _SCREAMING_SNAKE_CASE : List[Any]=0.1 , _SCREAMING_SNAKE_CASE : Tuple=1E-6 , _SCREAMING_SNAKE_CASE : int=1.0 , _SCREAMING_SNAKE_CASE : Optional[Any]="gelu_new" , _SCREAMING_SNAKE_CASE : Any=0 , _SCREAMING_SNAKE_CASE : int=False , _SCREAMING_SNAKE_CASE : Tuple=0 , _SCREAMING_SNAKE_CASE : int=1 , _SCREAMING_SNAKE_CASE : List[str]=False , _SCREAMING_SNAKE_CASE : List[str]=True , **_SCREAMING_SNAKE_CASE : Optional[Any] , ): """simple docstring""" UpperCamelCase = vocab_size UpperCamelCase = hidden_size UpperCamelCase = d_kv UpperCamelCase = d_ff UpperCamelCase = num_layers UpperCamelCase = num_heads UpperCamelCase = relative_attention_num_buckets UpperCamelCase = relative_attention_max_distance UpperCamelCase = dropout_rate UpperCamelCase = layer_norm_epsilon UpperCamelCase = initializer_factor UpperCamelCase = use_cache UpperCamelCase = eos_token_id UpperCamelCase = decoder_start_token_id # for backwards compatibility UpperCamelCase = dense_act_fn super().__init__( pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , tie_word_embeddings=snake_case_ , is_decoder=snake_case_ , **snake_case_ , ) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Optional[int] , _SCREAMING_SNAKE_CASE : List[str] , **_SCREAMING_SNAKE_CASE : Any ): """simple docstring""" cls._set_token_in_kwargs(snake_case_ ) UpperCamelCase , UpperCamelCase = cls.get_config_dict(snake_case_ , **snake_case_ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get('model_type' ) == "pix2struct": UpperCamelCase = config_dict['text_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case__ = """pix2struct_vision_model""" def __init__( self : str , _SCREAMING_SNAKE_CASE : Any=768 , _SCREAMING_SNAKE_CASE : Union[str, Any]=768 , _SCREAMING_SNAKE_CASE : Any=2048 , _SCREAMING_SNAKE_CASE : str=64 , _SCREAMING_SNAKE_CASE : Dict=12 , _SCREAMING_SNAKE_CASE : List[str]=12 , _SCREAMING_SNAKE_CASE : Any="gelu_new" , _SCREAMING_SNAKE_CASE : Tuple=1E-6 , _SCREAMING_SNAKE_CASE : int=0.0 , _SCREAMING_SNAKE_CASE : Optional[int]=0.0 , _SCREAMING_SNAKE_CASE : List[str]=1E-1_0 , _SCREAMING_SNAKE_CASE : int=1.0 , _SCREAMING_SNAKE_CASE : int=4096 , _SCREAMING_SNAKE_CASE : int=32 , _SCREAMING_SNAKE_CASE : Optional[int]=128 , **_SCREAMING_SNAKE_CASE : Union[str, Any] , ): """simple docstring""" super().__init__(**snake_case_ ) UpperCamelCase = hidden_size UpperCamelCase = patch_embed_hidden_size UpperCamelCase = d_ff UpperCamelCase = dropout_rate UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = initializer_range UpperCamelCase = initializer_factor UpperCamelCase = attention_dropout UpperCamelCase = layer_norm_eps UpperCamelCase = dense_act_fn UpperCamelCase = seq_len UpperCamelCase = relative_attention_num_buckets UpperCamelCase = relative_attention_max_distance UpperCamelCase = d_kv @classmethod def _SCREAMING_SNAKE_CASE ( cls : Optional[int] , _SCREAMING_SNAKE_CASE : int , **_SCREAMING_SNAKE_CASE : int ): """simple docstring""" cls._set_token_in_kwargs(snake_case_ ) UpperCamelCase , UpperCamelCase = cls.get_config_dict(snake_case_ , **snake_case_ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get('model_type' ) == "pix2struct": UpperCamelCase = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(snake_case_ , **snake_case_ ) class A__ ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case__ = """pix2struct""" snake_case__ = True def __init__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : List[str]=None , _SCREAMING_SNAKE_CASE : Union[str, Any]=None , _SCREAMING_SNAKE_CASE : Dict=1.0 , _SCREAMING_SNAKE_CASE : Tuple=0.0_2 , _SCREAMING_SNAKE_CASE : Optional[Any]=False , _SCREAMING_SNAKE_CASE : Optional[Any]=False , _SCREAMING_SNAKE_CASE : Tuple=True , **_SCREAMING_SNAKE_CASE : Optional[Any] , ): """simple docstring""" super().__init__(tie_word_embeddings=snake_case_ , is_encoder_decoder=snake_case_ , **snake_case_ ) if text_config is None: UpperCamelCase = {} logger.info('text_config is None. Initializing the Pix2StructTextConfig with default values.' ) if vision_config is None: UpperCamelCase = {} logger.info('vision_config is None. Initializing the Pix2StructVisionConfig with default values.' ) UpperCamelCase = PixaStructTextConfig(**snake_case_ ) UpperCamelCase = PixaStructVisionConfig(**snake_case_ ) UpperCamelCase = self.text_config.decoder_start_token_id UpperCamelCase = self.text_config.pad_token_id UpperCamelCase = self.text_config.eos_token_id UpperCamelCase = initializer_factor UpperCamelCase = initializer_range UpperCamelCase = self.initializer_range UpperCamelCase = self.initializer_range UpperCamelCase = is_vqa @classmethod def _SCREAMING_SNAKE_CASE ( cls : Dict , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Any , **_SCREAMING_SNAKE_CASE : Tuple ): """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" UpperCamelCase = copy.deepcopy(self.__dict__ ) UpperCamelCase = self.text_config.to_dict() UpperCamelCase = self.vision_config.to_dict() UpperCamelCase = self.__class__.model_type return output
280
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class __magic_name__ : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=6 , snake_case_=17 , snake_case_=23 , snake_case_=11 , snake_case_=True , ): lowercase =parent lowercase =batch_size lowercase =seq_length lowercase =act_dim lowercase =state_dim lowercase =hidden_size lowercase =max_length lowercase =is_training def _A( self ): lowercase =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =floats_tensor((self.batch_size, self.seq_length, 1) ) lowercase =ids_tensor((self.batch_size, self.seq_length) , vocab_size=10_00 ) lowercase =random_attention_mask((self.batch_size, self.seq_length) ) lowercase =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def _A( self ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def _A( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): lowercase =DecisionTransformerModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() lowercase =model(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def _A( self ): lowercase =self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) =config_and_inputs lowercase ={ '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class __magic_name__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = (DecisionTransformerModel,) if is_torch_available() else () UpperCamelCase__ = () UpperCamelCase__ = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids UpperCamelCase__ = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False def _A( self ): lowercase =DecisionTransformerModelTester(self ) lowercase =ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def _A( self ): self.config_tester.run_common_tests() def _A( self ): lowercase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @slow def _A( self ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase =DecisionTransformerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _A( self ): lowercase , lowercase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase =model_class(snake_case_ ) lowercase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase =[*signature.parameters.keys()] lowercase =[ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) @require_torch class __magic_name__ ( unittest.TestCase ): @slow def _A( self ): lowercase =2 # number of steps of autoregressive prediction we will perform lowercase =10 # defined by the RL environment, may be normalized lowercase =DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) lowercase =model.to(snake_case_ ) lowercase =model.config torch.manual_seed(0 ) lowercase =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ) # env.reset() lowercase =torch.tensor( [[0.24_27_93, -0.28_69_30_74, 0.8_74_26_13], [0.67_81_52_74, -0.08_10_10_85, -0.12_95_21_47]] , device=snake_case_ ) lowercase =torch.tensor(snake_case_ , device=snake_case_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowercase =state lowercase =torch.zeros(1 , 0 , config.act_dim , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.zeros(1 , 0 , device=snake_case_ , dtype=torch.floataa ) lowercase =torch.tensor(0 , device=snake_case_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case_ ): lowercase =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case_ )] , dim=1 ) lowercase =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case_ )] , dim=1 ) lowercase =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowercase , lowercase , lowercase =model( states=snake_case_ , actions=snake_case_ , rewards=snake_case_ , returns_to_go=snake_case_ , timesteps=snake_case_ , attention_mask=snake_case_ , return_dict=snake_case_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowercase , lowercase , lowercase , lowercase =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case_ , dtype=torch.floataa ), 1.0, False, {}, ) lowercase =action_pred[0, -1] lowercase =torch.cat([states, state] , dim=1 ) lowercase =returns_to_go[0, -1] - reward lowercase =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowercase =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case_ , dtype=torch.long ) * (step + 1)] , dim=1 )
72
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class _lowerCAmelCase ( metaclass=__SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE : str = ['note_seq'] def __init__(self , *lowercase , **lowercase ): requires_backends(self , ["""note_seq"""] ) @classmethod def _a (cls , *lowercase , **lowercase ): requires_backends(cls , ["""note_seq"""] ) @classmethod def _a (cls , *lowercase , **lowercase ): requires_backends(cls , ["""note_seq"""] )
667
'''simple docstring''' from math import pi, sqrt, tan def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase =(height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(lowercase_ , 2 ) * torus_radius * tube_radius def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase =(sidea + sidea + sidea) / 2 lowercase =sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def UpperCamelCase ( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def UpperCamelCase ( lowercase_ : float ) -> float: '''simple docstring''' if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def UpperCamelCase ( lowercase_ : float , lowercase_ : float ) -> float: '''simple docstring''' if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def UpperCamelCase ( lowercase_ : int , lowercase_ : float ) -> float: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F"""Rectangle: {area_rectangle(10, 20) = }""") print(F"""Square: {area_square(10) = }""") print(F"""Triangle: {area_triangle(10, 10) = }""") print(F"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(F"""Parallelogram: {area_parallelogram(10, 20) = }""") print(F"""Rhombus: {area_rhombus(10, 20) = }""") print(F"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(F"""Circle: {area_circle(20) = }""") print(F"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(F"""Cube: {surface_area_cube(20) = }""") print(F"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(F"""Sphere: {surface_area_sphere(20) = }""") print(F"""Hemisphere: {surface_area_hemisphere(20) = }""") print(F"""Cone: {surface_area_cone(10, 20) = }""") print(F"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(F"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(F"""Torus: {surface_area_torus(20, 10) = }""") print(F"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(F"""Square: {area_reg_polygon(4, 10) = }""") print(F"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
72
0
'''simple docstring''' def a_ ( UpperCamelCase_ , UpperCamelCase_ = 0 ): A_ = length or len(lowercase_ ) A_ = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: A_ , A_ = list_data[i + 1], list_data[i] A_ = True return list_data if not swapped else bubble_sort(lowercase_ , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
452
'''simple docstring''' import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class __magic_name__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): UpperCamelCase__ = BarthezTokenizer UpperCamelCase__ = BarthezTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = True def _A( self ): super().setUp() lowercase =BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=snake_case_ ) lowercase =tokenizer def _A( self ): lowercase ='''<pad>''' lowercase =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def _A( self ): lowercase =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(snake_case_ ) , 10_11_22 ) def _A( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_11_22 ) @require_torch def _A( self ): lowercase =['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] lowercase =[0, 57, 30_18, 7_03_07, 91, 2] lowercase =self.tokenizer( snake_case_ , max_length=len(snake_case_ ) , padding=snake_case_ , truncation=snake_case_ , return_tensors='''pt''' ) self.assertIsInstance(snake_case_ , snake_case_ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) lowercase =batch.input_ids.tolist()[0] self.assertListEqual(snake_case_ , snake_case_ ) def _A( self ): if not self.test_rust_tokenizer: return lowercase =self.get_tokenizer() lowercase =self.get_rust_tokenizer() lowercase ='''I was born in 92000, and this is falsé.''' lowercase =tokenizer.tokenize(snake_case_ ) lowercase =rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) lowercase =self.get_rust_tokenizer() lowercase =tokenizer.encode(snake_case_ ) lowercase =rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) @slow def _A( self ): # fmt: off lowercase ={'''input_ids''': [[0, 4_90, 1_43_28, 45_07, 3_54, 47, 4_36_69, 95, 25, 7_81_17, 2_02_15, 1_97_79, 1_90, 22, 4_00, 4, 3_53_43, 8_03_10, 6_03, 86, 2_49_37, 1_05, 3_34_38, 9_47_62, 1_96, 3_96_42, 7, 15, 1_59_33, 1_73, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_05_34, 87, 25, 66, 33_58, 1_96, 5_52_89, 8, 8_29_61, 81, 22_04, 7_52_03, 7, 15, 7_63, 1_29_56, 2_16, 1_78, 1_43_28, 95_95, 13_77, 6_96_93, 7, 4_48, 7_10_21, 1_96, 1_81_06, 14_37, 1_39_74, 1_08, 90_83, 4, 4_93_15, 7, 39, 86, 13_26, 27_93, 4_63_33, 4, 4_48, 1_96, 7_45_88, 7, 4_93_15, 7, 39, 21, 8_22, 3_84_70, 74, 21, 6_67_23, 6_24_80, 8, 2_20_50, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase =[ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=snake_case_ , )
72
0