code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> int: return int((input_a, input_a).count(0 ) != 0 ) def __UpperCAmelCase ( ) -> None: assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
16
"""simple docstring""" def __UpperCAmelCase ( __lowerCamelCase = 50 ) -> int: lowercase__ : int = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(F'''{solution() = }''')
16
1
from __future__ import annotations def UpperCAmelCase__ ( lowerCamelCase ): return [ord(lowerCamelCase ) - 96 for elem in plain] def UpperCAmelCase__ ( lowerCamelCase ): return "".join(chr(elem + 96 ) for elem in encoded ) def UpperCAmelCase__ ( ): lowercase :str = encode(input("-> " ).strip().lower() ) print("Encoded: ", lowerCamelCase ) print("Decoded:", decode(lowerCamelCase ) ) if __name__ == "__main__": main()
158
from collections.abc import Iterable from typing import Generic, TypeVar _UpperCAmelCase : Any = TypeVar("_T") class __lowerCAmelCase ( Generic[_T]): def __init__( self: Union[str, Any] , _lowerCAmelCase: Iterable[_T] | None = None ): lowercase :list[_T] = list(iterable or [] ) lowercase :list[_T] = [] def __len__( self: Dict ): return len(self._stacka ) + len(self._stacka ) def __repr__( self: List[Any] ): return F"Queue({tuple(self._stacka[::-1] + self._stacka )})" def SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _lowerCAmelCase: _T ): self._stacka.append(_lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self: Any ): lowercase :int = self._stacka.pop lowercase :List[Any] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("Queue is empty" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
158
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging snake_case__ : List[str] = logging.get_logger(__name__) snake_case__ : int = '▁' snake_case__ : str = {'vocab_file': 'sentencepiece.bpe.model'} snake_case__ : Tuple = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), } } snake_case__ : str = { 'facebook/mbart-large-en-ro': 1024, 'facebook/mbart-large-cc25': 1024, } # fmt: off snake_case__ : Any = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class A_ ( _lowerCamelCase ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = ["""input_ids""", """attention_mask"""] lowerCAmelCase__ = [] lowerCAmelCase__ = [] def __init__(self :Union[str, Any] , _UpperCamelCase :Optional[int] , _UpperCamelCase :Any="<s>" , _UpperCamelCase :int="</s>" , _UpperCamelCase :Any="</s>" , _UpperCamelCase :str="<s>" , _UpperCamelCase :str="<unk>" , _UpperCamelCase :Optional[int]="<pad>" , _UpperCamelCase :Optional[Any]="<mask>" , _UpperCamelCase :str=None , _UpperCamelCase :List[str]=None , _UpperCamelCase :Any=None , _UpperCamelCase :Optional[Dict[str, Any]] = None , _UpperCamelCase :Tuple=None , **_UpperCamelCase :int , )-> str: # Mask token behave like a normal word, i.e. include the space before it __A = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else mask_token __A = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_UpperCamelCase , eos_token=_UpperCamelCase , unk_token=_UpperCamelCase , sep_token=_UpperCamelCase , cls_token=_UpperCamelCase , pad_token=_UpperCamelCase , mask_token=_UpperCamelCase , tokenizer_file=_UpperCamelCase , src_lang=_UpperCamelCase , tgt_lang=_UpperCamelCase , additional_special_tokens=_UpperCamelCase , sp_model_kwargs=self.sp_model_kwargs , **_UpperCamelCase , ) __A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_UpperCamelCase ) ) __A = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token __A = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __A = 1 __A = len(self.sp_model ) __A = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(_UpperCamelCase ) } __A = {v: k for k, v in self.lang_code_to_id.items()} __A = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) __A = {v: k for k, v in self.fairseq_tokens_to_ids.items()} __A = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) __A = src_lang if src_lang is not None else '''en_XX''' __A = self.lang_code_to_id[self._src_lang] __A = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__(self :Optional[int] )-> Dict: __A = self.__dict__.copy() __A = None __A = self.sp_model.serialized_model_proto() return state def __setstate__(self :Dict , _UpperCamelCase :List[Any] )-> Optional[int]: __A = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __A = {} __A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def _lowerCAmelCase (self :Tuple )-> Union[str, Any]: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def _lowerCAmelCase (self :List[str] )-> str: return self._src_lang @src_lang.setter def _lowerCAmelCase (self :Any , _UpperCamelCase :str )-> None: __A = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _lowerCAmelCase (self :Optional[Any] , _UpperCamelCase :List[int] , _UpperCamelCase :Optional[List[int]] = None , _UpperCamelCase :bool = False )-> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_UpperCamelCase , token_ids_a=_UpperCamelCase , already_has_special_tokens=_UpperCamelCase ) __A = [1] * len(self.prefix_tokens ) __A = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(_UpperCamelCase )) + suffix_ones return prefix_ones + ([0] * len(_UpperCamelCase )) + ([0] * len(_UpperCamelCase )) + suffix_ones def _lowerCAmelCase (self :Tuple , _UpperCamelCase :List[int] , _UpperCamelCase :Optional[List[int]] = None )-> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _lowerCAmelCase (self :List[Any] , _UpperCamelCase :List[int] , _UpperCamelCase :Optional[List[int]] = None )-> List[int]: __A = [self.sep_token_id] __A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowerCAmelCase (self :List[Any] , _UpperCamelCase :Tuple , _UpperCamelCase :str , _UpperCamelCase :Optional[str] , _UpperCamelCase :Optional[str] , **_UpperCamelCase :int )-> Optional[int]: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) __A = src_lang __A = self(_UpperCamelCase , add_special_tokens=_UpperCamelCase , return_tensors=_UpperCamelCase , **_UpperCamelCase ) __A = self.convert_tokens_to_ids(_UpperCamelCase ) __A = tgt_lang_id return inputs def _lowerCAmelCase (self :str )-> List[str]: __A = {self.convert_ids_to_tokens(_UpperCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _lowerCAmelCase (self :int , _UpperCamelCase :str )-> List[str]: return self.sp_model.encode(_UpperCamelCase , out_type=_UpperCamelCase ) def _lowerCAmelCase (self :Dict , _UpperCamelCase :Optional[Any] )-> Dict: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __A = self.sp_model.PieceToId(_UpperCamelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _lowerCAmelCase (self :List[str] , _UpperCamelCase :List[str] )-> str: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _lowerCAmelCase (self :Optional[Any] , _UpperCamelCase :List[str] )-> Dict: __A = ''''''.join(_UpperCamelCase ).replace(_UpperCamelCase , ''' ''' ).strip() return out_string def _lowerCAmelCase (self :Optional[Any] , _UpperCamelCase :str , _UpperCamelCase :Optional[str] = None )-> Tuple[str]: if not os.path.isdir(_UpperCamelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __A = os.path.join( _UpperCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _UpperCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(_UpperCamelCase , '''wb''' ) as fi: __A = self.sp_model.serialized_model_proto() fi.write(_UpperCamelCase ) return (out_vocab_file,) def _lowerCAmelCase (self :Dict , _UpperCamelCase :List[str] , _UpperCamelCase :str = "en_XX" , _UpperCamelCase :Optional[List[str]] = None , _UpperCamelCase :str = "ro_RO" , **_UpperCamelCase :str , )-> BatchEncoding: __A = src_lang __A = tgt_lang return super().prepare_seqaseq_batch(_UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ) def _lowerCAmelCase (self :Any )-> Optional[Any]: return self.set_src_lang_special_tokens(self.src_lang ) def _lowerCAmelCase (self :Union[str, Any] )-> int: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def _lowerCAmelCase (self :List[str] , _UpperCamelCase :Any )-> None: __A = self.lang_code_to_id[src_lang] __A = [] __A = [self.eos_token_id, self.cur_lang_code] def _lowerCAmelCase (self :Any , _UpperCamelCase :str )-> None: __A = self.lang_code_to_id[lang] __A = [] __A = [self.eos_token_id, self.cur_lang_code]
117
import os from datetime import datetime as dt from github import Github snake_case__ : Union[str, Any] = [ 'good first issue', 'feature request', 'wip', ] def _a ( ) -> List[Any]: '''simple docstring''' __A = Github(os.environ['''GITHUB_TOKEN'''] ) __A = g.get_repo('''huggingface/accelerate''' ) __A = repo.get_issues(state='''open''' ) for issue in open_issues: __A = sorted([comment for comment in issue.get_comments()] , key=lambda lowerCamelCase : i.created_at , reverse=lowerCamelCase ) __A = comments[0] if len(lowerCamelCase ) > 0 else None __A = dt.utcnow() __A = (current_time - issue.updated_at).days __A = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state='''closed''' ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( '''This issue has been automatically marked as stale because it has not had ''' '''recent activity. If you think this still needs to be addressed ''' '''please comment on this thread.\n\nPlease note that issues that do not follow the ''' '''[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) ''' '''are likely to be ignored.''' ) if __name__ == "__main__": main()
117
1
"""simple docstring""" def __lowerCamelCase ( ) -> int: return [ a * b * (10_00 - a - b) for a in range(1 , 9_99 ) for b in range(a_ , 9_99 ) if (a * a + b * b == (10_00 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'{solution() = }')
239
"""simple docstring""" def __lowerCamelCase ( a_ : int , a_ : int ) -> int: return int((input_a, input_a).count(0 ) == 0 ) def __lowerCamelCase ( ) -> None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
239
1
from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class _lowercase : '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = PegasusConfig SCREAMING_SNAKE_CASE__ : Tuple = {} SCREAMING_SNAKE_CASE__ : Tuple = '''gelu''' def __init__( self :str , lowerCAmelCase__ :Dict , lowerCAmelCase__ :List[Any]=13 , lowerCAmelCase__ :Union[str, Any]=7 , lowerCAmelCase__ :Optional[Any]=True , lowerCAmelCase__ :Union[str, Any]=False , lowerCAmelCase__ :Tuple=99 , lowerCAmelCase__ :Optional[Any]=32 , lowerCAmelCase__ :Optional[int]=2 , lowerCAmelCase__ :Union[str, Any]=4 , lowerCAmelCase__ :List[str]=37 , lowerCAmelCase__ :Optional[Any]=0.1 , lowerCAmelCase__ :int=0.1 , lowerCAmelCase__ :str=40 , lowerCAmelCase__ :Dict=2 , lowerCAmelCase__ :List[str]=1 , lowerCAmelCase__ :str=0 , ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE : str = parent __SCREAMING_SNAKE_CASE : List[str] = batch_size __SCREAMING_SNAKE_CASE : Dict = seq_length __SCREAMING_SNAKE_CASE : int = is_training __SCREAMING_SNAKE_CASE : Any = use_labels __SCREAMING_SNAKE_CASE : Dict = vocab_size __SCREAMING_SNAKE_CASE : List[Any] = hidden_size __SCREAMING_SNAKE_CASE : Any = num_hidden_layers __SCREAMING_SNAKE_CASE : Optional[int] = num_attention_heads __SCREAMING_SNAKE_CASE : Optional[int] = intermediate_size __SCREAMING_SNAKE_CASE : Any = hidden_dropout_prob __SCREAMING_SNAKE_CASE : List[Any] = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE : List[str] = max_position_embeddings __SCREAMING_SNAKE_CASE : List[str] = eos_token_id __SCREAMING_SNAKE_CASE : List[str] = pad_token_id __SCREAMING_SNAKE_CASE : Optional[int] = bos_token_id def __magic_name__( self :str ) -> Dict: __SCREAMING_SNAKE_CASE : Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __SCREAMING_SNAKE_CASE : Dict = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __SCREAMING_SNAKE_CASE : str = tf.concat([input_ids, eos_tensor] , axis=1 ) __SCREAMING_SNAKE_CASE : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE : Tuple = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) __SCREAMING_SNAKE_CASE : Tuple = prepare_pegasus_inputs_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return config, inputs_dict def __magic_name__( self :List[str] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :List[Any] ) -> Tuple: __SCREAMING_SNAKE_CASE : List[Any] = TFPegasusModel(config=lowerCAmelCase__ ).get_decoder() __SCREAMING_SNAKE_CASE : List[Any] = inputs_dict['''input_ids'''] __SCREAMING_SNAKE_CASE : Union[str, Any] = input_ids[:1, :] __SCREAMING_SNAKE_CASE : Optional[Any] = inputs_dict['''attention_mask'''][:1, :] __SCREAMING_SNAKE_CASE : str = inputs_dict['''head_mask'''] __SCREAMING_SNAKE_CASE : Union[str, Any] = 1 # first forward pass __SCREAMING_SNAKE_CASE : Any = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , head_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __SCREAMING_SNAKE_CASE : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) __SCREAMING_SNAKE_CASE : Union[str, Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __SCREAMING_SNAKE_CASE : List[str] = tf.concat([input_ids, next_tokens] , axis=-1 ) __SCREAMING_SNAKE_CASE : Dict = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __SCREAMING_SNAKE_CASE : Optional[int] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ )[0] __SCREAMING_SNAKE_CASE : Optional[int] = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __SCREAMING_SNAKE_CASE : List[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __SCREAMING_SNAKE_CASE : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx] __SCREAMING_SNAKE_CASE : Union[str, Any] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase__ , lowerCAmelCase__ , rtol=1E-3 ) def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , ): if attention_mask is None: __SCREAMING_SNAKE_CASE : str = tf.cast(tf.math.not_equal(lowercase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __SCREAMING_SNAKE_CASE : str = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __SCREAMING_SNAKE_CASE : Optional[Any] = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __SCREAMING_SNAKE_CASE : str = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __SCREAMING_SNAKE_CASE : Tuple = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _lowercase ( A__ , A__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : List[str] = (TFPegasusForConditionalGeneration,) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : Union[str, Any] = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE__ : Dict = True SCREAMING_SNAKE_CASE__ : List[Any] = False SCREAMING_SNAKE_CASE__ : Optional[Any] = False def __magic_name__( self :Optional[Any] ) -> Any: __SCREAMING_SNAKE_CASE : List[str] = TFPegasusModelTester(self ) __SCREAMING_SNAKE_CASE : Any = ConfigTester(self , config_class=lowerCAmelCase__ ) def __magic_name__( self :Tuple ) -> List[Any]: self.config_tester.run_common_tests() def __magic_name__( self :Tuple ) -> Dict: __SCREAMING_SNAKE_CASE : Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase__ ) @require_sentencepiece @require_tokenizers @require_tf class _lowercase ( unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ''', ] SCREAMING_SNAKE_CASE__ : int = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re "grateful" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers SCREAMING_SNAKE_CASE__ : Optional[Any] = '''google/pegasus-xsum''' @cached_property def __magic_name__( self :Tuple ) -> Optional[Any]: return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def __magic_name__( self :List[str] ) -> Optional[Any]: __SCREAMING_SNAKE_CASE : List[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def __magic_name__( self :Union[str, Any] , **lowerCAmelCase__ :List[Any] ) -> Optional[int]: __SCREAMING_SNAKE_CASE : Union[str, Any] = self.translate_src_text(**lowerCAmelCase__ ) assert self.expected_text == generated_words def __magic_name__( self :Any , **lowerCAmelCase__ :int ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE : Tuple = self.tokenizer(self.src_text , **lowerCAmelCase__ , padding=lowerCAmelCase__ , return_tensors='''tf''' ) __SCREAMING_SNAKE_CASE : str = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE : Any = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=lowerCAmelCase__ ) return generated_words @slow def __magic_name__( self :Tuple ) -> int: self._assert_generated_batch_equal_expected()
9
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex a_ = logging.getLogger(__name__) class _UpperCamelCase : '''simple docstring''' def __init__( self : Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = False def __UpperCamelCase ( self : str , a : str , a : Optional[int] , a : Any , a : str ) -> List[Any]: """simple docstring""" if not self.initialized: SCREAMING_SNAKE_CASE : List[str] = RagRetriever( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[int] = True def __UpperCamelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" self.retriever.index.init_index() def __UpperCamelCase ( self : Optional[Any] , a : List[Any] , a : Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Dict = self.retriever._main_retrieve(a , a ) return doc_ids, retrieved_doc_embeds class _UpperCamelCase ( __A ): '''simple docstring''' def __init__( self : Tuple , a : Any , a : Tuple , a : Tuple , a : Tuple , a : List[Any]=None ) -> Optional[int]: """simple docstring""" if index is not None and index.is_initialized() and len(a ) > 0: raise ValueError( "When using Ray for distributed fine-tuning, " "you'll need to provide the paths instead, " "as the dataset and the index are loaded " "separately. More info in examples/rag/use_own_knowledge_dataset.py " ) super().__init__( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[Any] = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(a , a , a , a ) for worker in self.retrieval_workers ] ) def __UpperCamelCase ( self : Any ) -> Dict: """simple docstring""" logger.info("initializing retrieval" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def __UpperCamelCase ( self : Tuple , a : Optional[int] , a : Any ) -> int: """simple docstring""" if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. SCREAMING_SNAKE_CASE : Optional[Any] = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : str = ray.get(random_worker.retrieve.remote(a , a ) ) else: SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Any = self._main_retrieve(a , a ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(a ) @classmethod def __UpperCamelCase ( cls : str , a : Optional[Any] , a : Any=None , **a : List[Any] ) -> str: """simple docstring""" return super(a , cls ).get_tokenizers(a , a , **a ) @classmethod def __UpperCamelCase ( cls : Union[str, Any] , a : int , a : Any , a : List[Any]=None , **a : Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : str = kwargs.pop("config" , a ) or RagConfig.from_pretrained(a , **a ) SCREAMING_SNAKE_CASE : List[Any] = RagTokenizer.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.question_encoder SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.generator if indexed_dataset is not None: SCREAMING_SNAKE_CASE : str = "custom" SCREAMING_SNAKE_CASE : List[Any] = CustomHFIndex(config.retrieval_vector_size , a ) else: SCREAMING_SNAKE_CASE : List[str] = cls._build_index(a ) return cls( a , question_encoder_tokenizer=a , generator_tokenizer=a , retrieval_workers=a , index=a , )
76
0
"""simple docstring""" import os from datetime import datetime as dt from github import Github _a = [ 'good first issue', 'feature request', 'wip', ] def __a ( ): UpperCAmelCase_ : Optional[Any] = Github(os.environ["GITHUB_TOKEN"] ) UpperCAmelCase_ : List[Any] = g.get_repo("huggingface/accelerate" ) UpperCAmelCase_ : Union[str, Any] = repo.get_issues(state="open" ) for issue in open_issues: UpperCAmelCase_ : Union[str, Any] = sorted([comment for comment in issue.get_comments()], key=lambda __lowerCamelCase : i.created_at, reverse=__lowerCamelCase ) UpperCAmelCase_ : Any = comments[0] if len(__lowerCamelCase ) > 0 else None UpperCAmelCase_ : Optional[Any] = dt.utcnow() UpperCAmelCase_ : Optional[int] = (current_time - issue.updated_at).days UpperCAmelCase_ : str = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state="closed" ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) " "are likely to be ignored." ) if __name__ == "__main__": main()
23
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _a = 0 _a = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _a = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _a = tuple[int, int] class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" UpperCAmelCase_ : int = pos_x UpperCAmelCase_ : List[Any] = pos_y UpperCAmelCase_ : Union[str, Any] = (pos_y, pos_x) UpperCAmelCase_ : Any = goal_x UpperCAmelCase_ : Dict = goal_y UpperCAmelCase_ : Any = g_cost UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = self.calculate_heuristic() UpperCAmelCase_ : Any = self.g_cost + self.h_cost def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.pos_x - self.goal_x UpperCAmelCase_ : Union[str, Any] = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowercase_ ) + abs(lowercase_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowercase_ ): """simple docstring""" return self.f_cost < other.f_cost class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowercase_ ) UpperCAmelCase_ : List[Any] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowercase_ ) UpperCAmelCase_ : str = [self.start] UpperCAmelCase_ : list[Node] = [] UpperCAmelCase_ : int = False def UpperCamelCase__ ( self ): """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase_ : List[str] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowercase_ ) self.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : str = self.get_successors(lowercase_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : Union[str, Any] = self.open_nodes.pop(self.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowercase_ ) else: self.open_nodes.append(lowercase_ ) return [self.start.pos] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = [] for action in delta: UpperCAmelCase_ : str = parent.pos_x + action[1] UpperCAmelCase_ : int = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowercase_ , lowercase_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowercase_ , ) ) return successors def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = node UpperCAmelCase_ : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase_ : Optional[int] = current_node.parent path.reverse() return path class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Dict = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = False def UpperCamelCase__ ( self ): """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() UpperCAmelCase_ : List[str] = self.fwd_astar.open_nodes.pop(0 ) UpperCAmelCase_ : List[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowercase_ , lowercase_ ) self.fwd_astar.closed_nodes.append(lowercase_ ) self.bwd_astar.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : Tuple = current_bwd_node UpperCAmelCase_ : str = current_fwd_node UpperCAmelCase_ : Dict = { self.fwd_astar: self.fwd_astar.get_successors(lowercase_ ), self.bwd_astar: self.bwd_astar.get_successors(lowercase_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : List[Any] = astar.open_nodes.pop( astar.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowercase_ ) else: astar.open_nodes.append(lowercase_ ) return [self.fwd_astar.start.pos] def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.fwd_astar.retrace_path(lowercase_ ) UpperCAmelCase_ : int = self.bwd_astar.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() UpperCAmelCase_ : Any = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _a = (0, 0) _a = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _a = time.time() _a = AStar(init, goal) _a = a_star.search() _a = time.time() - start_time print(f"""AStar execution time = {end_time:f} seconds""") _a = time.time() _a = BidirectionalAStar(init, goal) _a = time.time() - bd_start_time print(f"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
23
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''google/canine-s''': '''https://huggingface.co/google/canine-s/resolve/main/config.json''', # See all CANINE models at https://huggingface.co/models?filter=canine } class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = '''canine''' def __init__( self : Dict , __UpperCAmelCase : List[str]=768 , __UpperCAmelCase : str=12 , __UpperCAmelCase : Union[str, Any]=12 , __UpperCAmelCase : int=3072 , __UpperCAmelCase : Optional[int]="gelu" , __UpperCAmelCase : Tuple=0.1 , __UpperCAmelCase : str=0.1 , __UpperCAmelCase : List[Any]=16384 , __UpperCAmelCase : Any=16 , __UpperCAmelCase : str=0.02 , __UpperCAmelCase : Dict=1E-12 , __UpperCAmelCase : Optional[Any]=0 , __UpperCAmelCase : int=0xE000 , __UpperCAmelCase : List[Any]=0xE001 , __UpperCAmelCase : Any=4 , __UpperCAmelCase : Dict=4 , __UpperCAmelCase : List[str]=8 , __UpperCAmelCase : int=16384 , __UpperCAmelCase : Union[str, Any]=128 , **__UpperCAmelCase : Dict , ): '''simple docstring''' super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) _A = max_position_embeddings _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = initializer_range _A = type_vocab_size _A = layer_norm_eps # Character config: _A = downsampling_rate _A = upsampling_kernel_size _A = num_hash_functions _A = num_hash_buckets _A = local_transformer_stride
79
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __UpperCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): lowercase : List[Any] =StableDiffusionInstructPixaPixPipeline lowercase : List[Any] =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width', 'cross_attention_kwargs'} lowercase : Optional[Any] =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowercase : Union[str, Any] =IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase : List[Any] =IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCamelCase_ =UNetaDConditionModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=8, out_channels=4, down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D'''), up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D'''), cross_attention_dim=32, ) lowerCamelCase_ =PNDMScheduler(skip_prk_steps=lowerCAmelCase ) torch.manual_seed(0 ) lowerCamelCase_ =AutoencoderKL( block_out_channels=[32, 64], in_channels=3, out_channels=3, down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], latent_channels=4, ) torch.manual_seed(0 ) lowerCamelCase_ =CLIPTextConfig( bos_token_id=0, eos_token_id=2, hidden_size=32, intermediate_size=37, layer_norm_eps=1e-05, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=1_000, ) lowerCamelCase_ =CLIPTextModel(lowerCAmelCase ) lowerCamelCase_ =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCamelCase_ ={ '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=0 ): """simple docstring""" lowerCamelCase_ =floats_tensor((1, 3, 32, 32), rng=random.Random(lowerCAmelCase ) ).to(lowerCAmelCase ) lowerCamelCase_ =image.cpu().permute(0, 2, 3, 1 )[0] lowerCamelCase_ =Image.fromarray(np.uinta(lowerCAmelCase ) ).convert('''RGB''' ) if str(lowerCAmelCase ).startswith('''mps''' ): lowerCamelCase_ =torch.manual_seed(lowerCAmelCase ) else: lowerCamelCase_ =torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) lowerCamelCase_ ={ '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''image_guidance_scale''': 1, '''output_type''': '''numpy''', } return inputs def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ =self.get_dummy_components() lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase ) lowerCamelCase_ =sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowerCamelCase_ =self.get_dummy_inputs(lowerCAmelCase ) lowerCamelCase_ =sd_pipe(**lowerCAmelCase ).images lowerCamelCase_ =image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCamelCase_ =np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ =self.get_dummy_components() lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase ) lowerCamelCase_ =sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowerCamelCase_ =self.get_dummy_inputs(lowerCAmelCase ) lowerCamelCase_ ='''french fries''' lowerCamelCase_ =sd_pipe(**lowerCAmelCase, negative_prompt=lowerCAmelCase ) lowerCamelCase_ =output.images lowerCamelCase_ =image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCamelCase_ =np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ =self.get_dummy_components() lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase ) lowerCamelCase_ =sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowerCamelCase_ =self.get_dummy_inputs(lowerCAmelCase ) lowerCamelCase_ =[inputs['''prompt''']] * 2 lowerCamelCase_ =np.array(inputs['''image'''] ).astype(np.floataa ) / 2_5_5.0 lowerCamelCase_ =torch.from_numpy(lowerCAmelCase ).unsqueeze(0 ).to(lowerCAmelCase ) lowerCamelCase_ =image / 2 + 0.5 lowerCamelCase_ =image.permute(0, 3, 1, 2 ) lowerCamelCase_ =image.repeat(2, 1, 1, 1 ) lowerCamelCase_ =sd_pipe(**lowerCAmelCase ).images lowerCamelCase_ =image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) lowerCamelCase_ =np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''cpu''' # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ =self.get_dummy_components() lowerCamelCase_ =EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5, beta_end=0.0_1_2, beta_schedule='''scaled_linear''' ) lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase ) lowerCamelCase_ =sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowerCamelCase_ =self.get_dummy_inputs(lowerCAmelCase ) lowerCamelCase_ =sd_pipe(**lowerCAmelCase ).images lowerCamelCase_ =image[0, -3:, -3:, -1] lowerCamelCase_ =[round(lowerCAmelCase, 4 ) for x in image_slice.flatten().tolist()] print(''','''.join([str(lowerCAmelCase ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) lowerCamelCase_ =np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowercase__ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.get_dummy_components() lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline(**lowerCAmelCase ) lowerCamelCase_ =VaeImageProcessor(do_resize=lowerCAmelCase, do_normalize=lowerCAmelCase ) lowerCamelCase_ =pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowerCamelCase_ =pipe(**self.get_dummy_inputs_by_type(lowerCAmelCase, input_image_type='''pt''' ) )[0] lowerCamelCase_ =components['''vae'''] lowerCamelCase_ =self.get_dummy_inputs_by_type(lowerCAmelCase, input_image_type='''pt''' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): lowerCamelCase_ =vae.encode(inputs[image_param] ).latent_dist.mode() lowerCamelCase_ =pipe(**lowerCAmelCase )[0] lowerCamelCase_ =np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCAmelCase, 1e-4, '''passing latents as image input generate different result from passing image''' ) @slow @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): def lowercase__ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self, lowerCAmelCase=0 ): """simple docstring""" lowerCamelCase_ =torch.manual_seed(lowerCAmelCase ) lowerCamelCase_ =load_image( '''https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg''' ) lowerCamelCase_ ={ '''prompt''': '''turn him into a cyborg''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''image_guidance_scale''': 1.0, '''output_type''': '''numpy''', } return inputs def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''', safety_checker=lowerCAmelCase ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing() lowerCamelCase_ =self.get_inputs() lowerCamelCase_ =pipe(**lowerCAmelCase ).images lowerCamelCase_ =image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) lowerCamelCase_ =np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''', safety_checker=lowerCAmelCase ) lowerCamelCase_ =LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing() lowerCamelCase_ =self.get_inputs() lowerCamelCase_ =pipe(**lowerCAmelCase ).images lowerCamelCase_ =image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) lowerCamelCase_ =np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''', safety_checker=lowerCAmelCase ) lowerCamelCase_ =DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing() lowerCamelCase_ =self.get_inputs() lowerCamelCase_ =pipe(**lowerCAmelCase ).images lowerCamelCase_ =image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) lowerCamelCase_ =np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =0 def callback_fn(lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ) -> None: lowerCamelCase_ =True nonlocal number_of_steps number_of_steps += 1 if step == 1: lowerCamelCase_ =latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) lowerCamelCase_ =latents[0, -3:, -3:, -1] lowerCamelCase_ =np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: lowerCamelCase_ =latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) lowerCamelCase_ =latents[0, -3:, -3:, -1] lowerCamelCase_ =np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 lowerCamelCase_ =False lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''', safety_checker=lowerCAmelCase, torch_dtype=torch.floataa ) lowerCamelCase_ =pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing() lowerCamelCase_ =self.get_inputs() pipe(**lowerCAmelCase, callback=lowerCAmelCase, callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase__ ( self ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''', safety_checker=lowerCAmelCase, torch_dtype=torch.floataa ) lowerCamelCase_ =pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowerCamelCase_ =self.get_inputs() lowerCamelCase_ =pipe(**lowerCAmelCase ) lowerCamelCase_ =torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 lowerCamelCase_ =inputs['''image'''].resize((504, 504) ) lowerCamelCase_ ='''timbrooks/instruct-pix2pix''' lowerCamelCase_ =StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCAmelCase, safety_checker=lowerCAmelCase, ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing() lowerCamelCase_ =pipe(**lowerCAmelCase ) lowerCamelCase_ =output.images[0] lowerCamelCase_ =image[255:258, 383:386, -1] assert image.shape == (504, 504, 3) lowerCamelCase_ =np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
75
0
'''simple docstring''' import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def a ( __a , __a , __a , __a , __a ) -> Tuple: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = TapasConfig.from_json_file(__a ) # set absolute/relative position embeddings parameter UpperCamelCase__ :Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": UpperCamelCase__ :List[str] = TapasForQuestionAnswering(config=__a ) elif task == "WTQ": # run_task_main.py hparams UpperCamelCase__ :Dict = 4 UpperCamelCase__ :List[Any] = True # hparam_utils.py hparams UpperCamelCase__ :Any = 0.6_6_4_6_9_4 UpperCamelCase__ :List[Any] = 0.2_0_7_9_5_1 UpperCamelCase__ :Tuple = 0.1_2_1_1_9_4 UpperCamelCase__ :int = True UpperCamelCase__ :List[str] = True UpperCamelCase__ :Dict = False UpperCamelCase__ :Dict = 0.0_3_5_2_5_1_3 UpperCamelCase__ :List[str] = TapasForQuestionAnswering(config=__a ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams UpperCamelCase__ :Optional[int] = 4 UpperCamelCase__ :Optional[int] = False # hparam_utils.py hparams UpperCamelCase__ :Optional[Any] = 3_6.4_5_1_9 UpperCamelCase__ :Tuple = 0.9_0_3_4_2_1 UpperCamelCase__ :Union[str, Any] = 2_2_2.0_8_8 UpperCamelCase__ :Any = True UpperCamelCase__ :Tuple = True UpperCamelCase__ :str = True UpperCamelCase__ :Any = 0.7_6_3_1_4_1 UpperCamelCase__ :List[str] = TapasForQuestionAnswering(config=__a ) elif task == "TABFACT": UpperCamelCase__ :List[Any] = TapasForSequenceClassification(config=__a ) elif task == "MLM": UpperCamelCase__ :Optional[int] = TapasForMaskedLM(config=__a ) elif task == "INTERMEDIATE_PRETRAINING": UpperCamelCase__ :List[str] = TapasModel(config=__a ) else: raise ValueError(f'''Task {task} not supported.''' ) print(f'''Building PyTorch model from configuration: {config}''' ) # Load weights from tf checkpoint load_tf_weights_in_tapas(__a , __a , __a ) # Save pytorch-model (weights and configuration) print(f'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__a ) # Save tokenizer files print(f'''Save tokenizer files to {pytorch_dump_path}''' ) UpperCamelCase__ :Optional[int] = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + '''vocab.txt''' , model_max_length=512 ) tokenizer.save_pretrained(__a ) print('''Used relative position embeddings:''' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''SQA''', type=str, help='''Model task for which to convert a checkpoint. Defaults to SQA.''' ) parser.add_argument( '''--reset_position_index_per_cell''', default=False, action='''store_true''', help='''Whether to use relative position embeddings or not. Defaults to True.''', ) parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--tapas_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained TAPAS model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __snake_case = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
219
'''simple docstring''' import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = {'''vocab_file''': '''vocab.txt'''} __snake_case = { '''vocab_file''': { '''openbmb/cpm-ant-10b''': '''https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt''', }, } __snake_case = { '''openbmb/cpm-ant-10b''': 1024, } def a ( __a ) -> Tuple: '''simple docstring''' UpperCamelCase__ :List[str] = collections.OrderedDict() with open(__a , '''r''' , encoding='''utf-8''' ) as reader: UpperCamelCase__ :Dict = reader.readlines() for index, token in enumerate(__a ): UpperCamelCase__ :str = token.rstrip('''\n''' ) UpperCamelCase__ :Optional[int] = index return vocab class lowercase ( A__ ): """simple docstring""" def __init__( self , UpperCamelCase_ , UpperCamelCase_="<unk>" , UpperCamelCase_=200 ): '''simple docstring''' UpperCamelCase__ :Tuple = vocab UpperCamelCase__ :List[str] = unk_token UpperCamelCase__ :Tuple = max_input_chars_per_word def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Optional[int] = list(UpperCamelCase_ ) if len(UpperCamelCase_ ) > self.max_input_chars_per_word: return [self.unk_token] UpperCamelCase__ :List[Any] = 0 UpperCamelCase__ :str = [] while start < len(UpperCamelCase_ ): UpperCamelCase__ :int = len(UpperCamelCase_ ) UpperCamelCase__ :List[Any] = None while start < end: UpperCamelCase__ :int = ''''''.join(chars[start:end] ) if substr in self.vocab: UpperCamelCase__ :List[Any] = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(UpperCamelCase_ ) UpperCamelCase__ :Any = end return sub_tokens class lowercase ( A__ ): """simple docstring""" _a = VOCAB_FILES_NAMES _a = PRETRAINED_VOCAB_FILES_MAP _a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a = ['input_ids', 'attention_mask'] _a = False def __init__( self , UpperCamelCase_ , UpperCamelCase_="<d>" , UpperCamelCase_="</d>" , UpperCamelCase_="<s>" , UpperCamelCase_="</s>" , UpperCamelCase_="<pad>" , UpperCamelCase_="<unk>" , UpperCamelCase_="</n>" , UpperCamelCase_="</_>" , UpperCamelCase_="left" , **UpperCamelCase_ , ): '''simple docstring''' requires_backends(self , ['''jieba'''] ) super().__init__( bod_token=UpperCamelCase_ , eod_token=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , line_token=UpperCamelCase_ , space_token=UpperCamelCase_ , padding_side=UpperCamelCase_ , **UpperCamelCase_ , ) UpperCamelCase__ :Tuple = bod_token UpperCamelCase__ :Dict = eod_token UpperCamelCase__ :Optional[int] = load_vocab(UpperCamelCase_ ) UpperCamelCase__ :Tuple = self.encoder[space_token] UpperCamelCase__ :List[Any] = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] UpperCamelCase__ :Union[str, Any] = collections.OrderedDict(sorted(self.encoder.items() , key=lambda UpperCamelCase_ : x[1] ) ) UpperCamelCase__ :Union[str, Any] = {v: k for k, v in self.encoder.items()} UpperCamelCase__ :List[Any] = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def lowerCAmelCase__ ( self ): '''simple docstring''' return self.encoder[self.bod_token] @property def lowerCAmelCase__ ( self ): '''simple docstring''' return self.encoder[self.eod_token] @property def lowerCAmelCase__ ( self ): '''simple docstring''' return self.encoder["\n"] @property def lowerCAmelCase__ ( self ): '''simple docstring''' return len(self.encoder ) def lowerCAmelCase__ ( self ): '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[Any] = [] for x in jieba.cut(UpperCamelCase_ , cut_all=UpperCamelCase_ ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(UpperCamelCase_ ) ) return output_tokens def lowerCAmelCase__ ( self , UpperCamelCase_ , **UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[Any] = [i for i in token_ids if i >= 0] UpperCamelCase__ :Optional[int] = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return token in self.encoder def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return "".join(UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return self.encoder.get(UpperCamelCase_ , self.encoder.get(self.unk_token ) ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return self.decoder.get(UpperCamelCase_ , self.unk_token ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' if os.path.isdir(UpperCamelCase_ ): UpperCamelCase__ :int = os.path.join( UpperCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) else: UpperCamelCase__ :str = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory UpperCamelCase__ :Any = 0 if " " in self.encoder: UpperCamelCase__ :Dict = self.encoder[''' '''] del self.encoder[" "] if "\n" in self.encoder: UpperCamelCase__ :List[str] = self.encoder['''\n'''] del self.encoder["\n"] UpperCamelCase__ :List[str] = collections.OrderedDict(sorted(self.encoder.items() , key=lambda UpperCamelCase_ : x[1] ) ) with open(UpperCamelCase_ , '''w''' , encoding='''utf-8''' ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( F'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' ''' Please check that the vocabulary is not corrupted!''' ) UpperCamelCase__ :Any = token_index writer.write(token + '''\n''' ) index += 1 return (vocab_file,) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_ ) if token_ids_a is not None: return [1] + ([0] * len(UpperCamelCase_ )) + [1] + ([0] * len(UpperCamelCase_ )) return [1] + ([0] * len(UpperCamelCase_ ))
219
1
'''simple docstring''' from collections.abc import Iterable from typing import Generic, TypeVar _SCREAMING_SNAKE_CASE = TypeVar("_T") class lowerCAmelCase_ ( Generic[_T] ): def __init__( self , _lowerCAmelCase = None ) -> None: _lowerCAmelCase = list(iterable or [] ) _lowerCAmelCase = [] def __len__( self ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self ) -> str: return f'''Queue({tuple(self._stacka[::-1] + self._stacka )})''' def _snake_case ( self , _lowerCAmelCase ) -> None: self._stacka.append(_lowerCAmelCase ) def _snake_case ( self ) -> _T: _lowerCAmelCase = self._stacka.pop _lowerCAmelCase = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("Queue is empty" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
158
'''simple docstring''' def __a(SCREAMING_SNAKE_CASE_ : int = 1000 ): '''simple docstring''' return sum(e for e in range(3 , SCREAMING_SNAKE_CASE_ ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(f'''{solution() = }''')
158
1
import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def __UpperCAmelCase ( __a : int ) -> List[str]: """simple docstring""" _a : List[str] = os.path.join(args.tf_model_dir ,'''parameters.json''' ) _a : Optional[int] = json.loads(open(lowerCamelCase_ ).read() ) if not params: raise ValueError( F"""It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.""" ) if not args.output.endswith('''.pt''' ): _a : List[str] = args.output + '.pt' _a : Tuple = OrderedDict() with tf.device('''/CPU:0''' ): _a : Any = tf.train.load_checkpoint(args.tf_model_dir ) _a : List[str] = reader.get_variable_to_shape_map() for key_name in shapes.keys(): _a : Union[str, Any] = reader.get_tensor(lowerCamelCase_ ).astype(np.floataa ) if key_name.endswith('''/adam_m''' ) or key_name.endswith('''/adam_v''' ): continue if key_name.startswith('''pasts/''' ): if key_name.startswith('''pasts/mlp''' ): _a : Optional[int] = int(key_name[9] ) elif key_name.startswith('''pasts/out''' ): _a : int = 8 _a : List[Any] = 'model.sqout.%d.weight' % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time _a : Optional[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _a : List[Any] = torch.tensor(lowerCamelCase_ ) elif key_name.startswith('''model/moe''' ): _a : List[str] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/switch_gating/kernel''' ): _a : Union[str, Any] = 'model.blocks.%d.feed_forward.mlp.router.classifier.weight' % player _a : Optional[int] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _a : Any = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/softmlp/kernel''' ): _a : Any = 'model.blocks.%d.feed_forward.soft_bypass_mlp.weight' % player _a : List[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _a : List[str] = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/wo/kernel''' ) or key_name.endswith('''/wi/kernel''' ): _a : List[Any] = key_name[-9:-7] for i in range(16 ): _a : Tuple = 'model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight' % (player, i, nlayer) _a : List[str] = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided _a : Optional[int] = torch.tensor(lowerCamelCase_ ) elif key_name.startswith('''model/mlp''' ): _a : Optional[int] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/p1/kernel''' ): _a : Dict = 'model.blocks.%d.feed_forward.mlp.wi.weight' % player _a : Optional[Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _a : Any = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/p1/bias''' ): _a : Optional[int] = 'model.blocks.%d.feed_forward.mlp.wi.bias' % player _a : int = vnp.copy() # same because it is one dimensional _a : Optional[int] = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/p2/kernel''' ): _a : List[str] = 'model.blocks.%d.feed_forward.mlp.wo.weight' % player _a : Union[str, Any] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _a : Union[str, Any] = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/p2/bias''' ): _a : Tuple = 'model.blocks.%d.feed_forward.mlp.wo.bias' % player _a : Optional[int] = vnp.copy() # same because it is one dimensional _a : int = torch.tensor(lowerCamelCase_ ) elif key_name.startswith('''model/ln''' ): _a : Union[str, Any] = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): _a : List[str] = 'model.blocks.%d.feed_forward.norm.bias' % player _a : Optional[Any] = vnp.copy() # same because it is one dimensional _a : Dict = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/g''' ): _a : Tuple = 'model.blocks.%d.feed_forward.norm.weight' % player _a : str = vnp.copy() # same because it is one dimensional _a : str = torch.tensor(lowerCamelCase_ ) elif key_name.startswith('''model/att''' ): _a : Dict = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/qkv/kernel''' ): _a : Optional[int] = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum _a : Union[str, Any] = state[:, 0, :, :] _a : str = state[:, 1, :, :] _a : Any = state[:, 2, :, :] _a : Optional[int] = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _a : Union[str, Any] = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _a : Optional[int] = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix _a : Optional[Any] = 'model.blocks.%d.self_attn.self_attn.q_proj.weight' % player _a : Any = torch.tensor(lowerCamelCase_ ) _a : Tuple = 'model.blocks.%d.self_attn.self_attn.k_proj.weight' % player _a : Optional[int] = torch.tensor(lowerCamelCase_ ) _a : Optional[int] = 'model.blocks.%d.self_attn.self_attn.v_proj.weight' % player _a : Dict = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/o/kernel''' ): _a : Optional[Any] = 'model.blocks.%d.self_attn.self_attn.out_proj.weight' % player _a : Tuple = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix _a : Optional[Any] = torch.tensor(lowerCamelCase_ ) elif key_name.startswith('''model/an''' ): _a : Tuple = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): _a : Optional[Any] = 'model.blocks.%d.self_attn.norm.bias' % player _a : Dict = vnp.copy() # same because it is one dimensional _a : Optional[int] = torch.tensor(lowerCamelCase_ ) elif key_name.endswith('''/g''' ): _a : Optional[Any] = 'model.blocks.%d.self_attn.norm.weight' % player _a : int = vnp.copy() # same because it is one dimensional _a : List[str] = torch.tensor(lowerCamelCase_ ) elif ( key_name.startswith('''model/wte''' ) or key_name.startswith('''model/wpe''' ) or key_name.startswith('''model/ete''' ) ): _a : Optional[int] = {'wte': 'embed_tokens', 'wpe': 'position_embeddings', 'ete': 'extra_position_embeddings'}[ key_name[-3:] ] _a : str = 'model.%s.weight' % nlayer _a : Optional[Any] = vnp.copy() # same in embedded _a : List[str] = torch.tensor(lowerCamelCase_ ) if key_name.startswith('''model/wte''' ): _a : Dict = 'lm_head.weight' _a : Optional[int] = vnp.copy() # same in embedded _a : Tuple = torch.tensor(lowerCamelCase_ ) elif key_name.startswith('''model/wob''' ): _a : Union[str, Any] = 'final_logits_bias' _a : Tuple = vnp.copy() # same in embedded _a : Union[str, Any] = state.reshape((1, -1) ) _a : List[Any] = torch.tensor(lowerCamelCase_ ) elif key_name == "model/dense/kernel": _a : Dict = 'model.last_project.weight' _a : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix _a : Optional[int] = torch.tensor(lowerCamelCase_ ) elif key_name == "model/dense_1/bias": _a : Optional[Any] = 'model.last_project.bias' _a : Optional[Any] = vnp.copy() # same because it is one dimensional _a : Optional[int] = torch.tensor(lowerCamelCase_ ) torch.save(lowerCamelCase_ ,args.output ) if __name__ == "__main__": a__ = argparse.ArgumentParser( description='''model converter.''', formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument('''--tf_model_dir''', metavar='''PATH''', type=str, required=True, help='''import model''') parser.add_argument('''--output''', metavar='''PATH''', type=str, required=True, help='''output model''') a__ = parser.parse_args() convert_tf_gptsan_to_pt(args)
367
import argparse import os import torch from transformers.utils import WEIGHTS_NAME a__ = ['''small''', '''medium''', '''large'''] a__ = '''lm_head.decoder.weight''' a__ = '''lm_head.weight''' def __UpperCAmelCase ( __a : str ,__a : str ) -> List[str]: """simple docstring""" _a : Any = torch.load(__a ) _a : List[str] = d.pop(__a ) os.makedirs(__a ,exist_ok=__a ) torch.save(__a ,os.path.join(__a ,__a ) ) if __name__ == "__main__": a__ = argparse.ArgumentParser() parser.add_argument('''--dialogpt_path''', default='''.''', type=str) a__ = parser.parse_args() for MODEL in DIALOGPT_MODELS: a__ = os.path.join(args.dialogpt_path, f'''{MODEL}_ft.pkl''') a__ = f'''./DialoGPT-{MODEL}''' convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
15
0
'''simple docstring''' import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class __magic_name__ ( unittest.TestCase, _UpperCAmelCase): def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : List[str] = load_tool("""text-to-speech""" ) self.tool.setup() def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): # SpeechT5 isn't deterministic torch.manual_seed(0 ) lowercase_ : List[str] = self.tool("""hey""" ) lowercase_ : Optional[int] = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_00_59_66_66_88_32_11_58_29, -0.0_00_36_57_64_01_90_79_50_64, -0.00_01_34_39_50_27_99_88_34_85] ) , ) ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): # SpeechT5 isn't deterministic torch.manual_seed(0 ) lowercase_ : Union[str, Any] = self.tool("""hey""" ) lowercase_ : int = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_00_59_66_66_88_32_11_58_29, -0.0_00_36_57_64_01_90_79_50_64, -0.00_01_34_39_50_27_99_88_34_85] ) , ) )
239
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase : Dict = {"configuration_mmbt": ["MMBTConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Dict = ["MMBTForClassification", "MMBTModel", "ModalEmbeddings"] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys _lowercase : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
239
1
import argparse import os from . import ( ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BART_PRETRAINED_MODEL_ARCHIVE_LIST, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, AlbertConfig, BartConfig, BertConfig, CamembertConfig, CTRLConfig, DistilBertConfig, DPRConfig, ElectraConfig, FlaubertConfig, GPTaConfig, LayoutLMConfig, LxmertConfig, OpenAIGPTConfig, RobertaConfig, TaConfig, TFAlbertForPreTraining, TFBartForConditionalGeneration, TFBartForSequenceClassification, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFCamembertForMaskedLM, TFCTRLLMHeadModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, TFElectraForPreTraining, TFFlaubertWithLMHeadModel, TFGPTaLMHeadModel, TFLayoutLMForMaskedLM, TFLxmertForPreTraining, TFLxmertVisualFeatureEncoder, TFOpenAIGPTLMHeadModel, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForSequenceClassification, TFTaForConditionalGeneration, TFTransfoXLLMHeadModel, TFWavaVecaModel, TFXLMRobertaForMaskedLM, TFXLMWithLMHeadModel, TFXLNetLMHeadModel, TransfoXLConfig, WavaVecaConfig, WavaVecaModel, XLMConfig, XLMRobertaConfig, XLNetConfig, is_torch_available, load_pytorch_checkpoint_in_tfa_model, ) from .utils import CONFIG_NAME, WEIGHTS_NAME, cached_file, logging if is_torch_available(): import numpy as np import torch from . import ( AlbertForPreTraining, BartForConditionalGeneration, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, CamembertForMaskedLM, CTRLLMHeadModel, DistilBertForMaskedLM, DistilBertForQuestionAnswering, DPRContextEncoder, DPRQuestionEncoder, DPRReader, ElectraForPreTraining, FlaubertWithLMHeadModel, GPTaLMHeadModel, LayoutLMForMaskedLM, LxmertForPreTraining, LxmertVisualFeatureEncoder, OpenAIGPTLMHeadModel, RobertaForMaskedLM, RobertaForSequenceClassification, TaForConditionalGeneration, TransfoXLLMHeadModel, XLMRobertaForMaskedLM, XLMWithLMHeadModel, XLNetLMHeadModel, ) logging.set_verbosity_info() _UpperCAmelCase : Any = { "bart": ( BartConfig, TFBartForConditionalGeneration, TFBartForSequenceClassification, BartForConditionalGeneration, BART_PRETRAINED_MODEL_ARCHIVE_LIST, ), "bert": ( BertConfig, TFBertForPreTraining, BertForPreTraining, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "bert-large-uncased-whole-word-masking-finetuned-squad": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "bert-large-cased-whole-word-masking-finetuned-squad": ( BertConfig, TFBertForQuestionAnswering, BertForQuestionAnswering, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "bert-base-cased-finetuned-mrpc": ( BertConfig, TFBertForSequenceClassification, BertForSequenceClassification, BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "dpr": ( DPRConfig, TFDPRQuestionEncoder, TFDPRContextEncoder, TFDPRReader, DPRQuestionEncoder, DPRContextEncoder, DPRReader, DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, ), "gpt2": ( GPTaConfig, TFGPTaLMHeadModel, GPTaLMHeadModel, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "xlnet": ( XLNetConfig, TFXLNetLMHeadModel, XLNetLMHeadModel, XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "xlm": ( XLMConfig, TFXLMWithLMHeadModel, XLMWithLMHeadModel, XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "xlm-roberta": ( XLMRobertaConfig, TFXLMRobertaForMaskedLM, XLMRobertaForMaskedLM, XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "transfo-xl": ( TransfoXLConfig, TFTransfoXLLMHeadModel, TransfoXLLMHeadModel, TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "openai-gpt": ( OpenAIGPTConfig, TFOpenAIGPTLMHeadModel, OpenAIGPTLMHeadModel, OPENAI_GPT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "roberta": ( RobertaConfig, TFRobertaForCausalLM, TFRobertaForMaskedLM, RobertaForMaskedLM, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "layoutlm": ( LayoutLMConfig, TFLayoutLMForMaskedLM, LayoutLMForMaskedLM, LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, ), "roberta-large-mnli": ( RobertaConfig, TFRobertaForSequenceClassification, RobertaForSequenceClassification, ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "camembert": ( CamembertConfig, TFCamembertForMaskedLM, CamembertForMaskedLM, CAMEMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "flaubert": ( FlaubertConfig, TFFlaubertWithLMHeadModel, FlaubertWithLMHeadModel, FLAUBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "distilbert": ( DistilBertConfig, TFDistilBertForMaskedLM, DistilBertForMaskedLM, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "distilbert-base-distilled-squad": ( DistilBertConfig, TFDistilBertForQuestionAnswering, DistilBertForQuestionAnswering, DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "lxmert": ( LxmertConfig, TFLxmertForPreTraining, LxmertForPreTraining, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "lxmert-visual-feature-encoder": ( LxmertConfig, TFLxmertVisualFeatureEncoder, LxmertVisualFeatureEncoder, LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "ctrl": ( CTRLConfig, TFCTRLLMHeadModel, CTRLLMHeadModel, CTRL_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "albert": ( AlbertConfig, TFAlbertForPreTraining, AlbertForPreTraining, ALBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "t5": ( TaConfig, TFTaForConditionalGeneration, TaForConditionalGeneration, T5_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "electra": ( ElectraConfig, TFElectraForPreTraining, ElectraForPreTraining, ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ), "wav2vec2": ( WavaVecaConfig, TFWavaVecaModel, WavaVecaModel, WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, ), } def A ( lowercase , lowercase , lowercase , lowercase , lowercase=False , lowercase=True ) -> Union[str, Any]: '''simple docstring''' if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type, should be one of {list(MODEL_CLASSES.keys() )}.''' ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = MODEL_CLASSES[model_type] # Initialise TF model if config_file in aws_config_map: UpperCamelCase = cached_file(lowercase , lowercase , force_download=not use_cached_models ) UpperCamelCase = config_class.from_json_file(lowercase ) UpperCamelCase = True UpperCamelCase = True print(f'''Building TensorFlow model from configuration: {config}''' ) UpperCamelCase = model_class(lowercase ) # Load weights from tf checkpoint if pytorch_checkpoint_path in aws_config_map.keys(): UpperCamelCase = cached_file( lowercase , lowercase , force_download=not use_cached_models ) # Load PyTorch checkpoint in tf2 model: UpperCamelCase = load_pytorch_checkpoint_in_tfa_model(lowercase , lowercase ) if compare_with_pt_model: UpperCamelCase = tf_model(tf_model.dummy_inputs , training=lowercase ) # build the network UpperCamelCase = torch.load(lowercase , map_location='cpu' ) UpperCamelCase = pt_model_class.from_pretrained( pretrained_model_name_or_path=lowercase , config=lowercase , state_dict=lowercase ) with torch.no_grad(): UpperCamelCase = pt_model(**pt_model.dummy_inputs ) UpperCamelCase = pto[0].numpy() UpperCamelCase = tfo[0].numpy() UpperCamelCase = np.amax(np.abs(np_pt - np_tf ) ) print(f'''Max absolute difference between models outputs {diff}''' ) assert diff <= 2e-2, f'''Error, model absolute difference is >2e-2: {diff}''' # Save pytorch-model print(f'''Save TensorFlow model to {tf_dump_path}''' ) tf_model.save_weights(lowercase , save_format='h5' ) def A ( lowercase , lowercase , lowercase=None , lowercase=None , lowercase=False , lowercase=False , lowercase=False , lowercase=False , ) -> Union[str, Any]: '''simple docstring''' if args_model_type is None: UpperCamelCase = list(MODEL_CLASSES.keys() ) else: UpperCamelCase = [args_model_type] for j, model_type in enumerate(lowercase , start=1 ): print('=' * 100 ) print(f''' Converting model type {j}/{len(lowercase )}: {model_type}''' ) print('=' * 100 ) if model_type not in MODEL_CLASSES: raise ValueError(f'''Unrecognized model type {model_type}, should be one of {list(MODEL_CLASSES.keys() )}.''' ) UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = MODEL_CLASSES[model_type] if model_shortcut_names_or_path is None: UpperCamelCase = list(aws_model_maps.keys() ) if config_shortcut_names_or_path is None: UpperCamelCase = model_shortcut_names_or_path for i, (model_shortcut_name, config_shortcut_name) in enumerate( zip(lowercase , lowercase ) , start=1 ): print('-' * 100 ) if "-squad" in model_shortcut_name or "-mrpc" in model_shortcut_name or "-mnli" in model_shortcut_name: if not only_convert_finetuned_models: print(f''' Skipping finetuned checkpoint {model_shortcut_name}''' ) continue UpperCamelCase = model_shortcut_name elif only_convert_finetuned_models: print(f''' Skipping not finetuned checkpoint {model_shortcut_name}''' ) continue print( f''' Converting checkpoint {i}/{len(lowercase )}: {model_shortcut_name} - model_type {model_type}''' ) print('-' * 100 ) if config_shortcut_name in aws_config_map: UpperCamelCase = cached_file(lowercase , lowercase , force_download=not use_cached_models ) else: UpperCamelCase = config_shortcut_name if model_shortcut_name in aws_model_maps: UpperCamelCase = cached_file(lowercase , lowercase , force_download=not use_cached_models ) else: UpperCamelCase = model_shortcut_name if os.path.isfile(lowercase ): UpperCamelCase = 'converted_model' convert_pt_checkpoint_to_tf( model_type=lowercase , pytorch_checkpoint_path=lowercase , config_file=lowercase , tf_dump_path=os.path.join(lowercase , model_shortcut_name + '-tf_model.h5' ) , compare_with_pt_model=lowercase , ) if remove_cached_files: os.remove(lowercase ) os.remove(lowercase ) if __name__ == "__main__": _UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_dump_path", default=None, type=str, required=True, help="Path to the output Tensorflow dump file." ) parser.add_argument( "--model_type", default=None, type=str, help=( F'''Model type selected in the list of {list(MODEL_CLASSES.keys())}. If not given, will download and ''' "convert all the models from AWS." ), ) parser.add_argument( "--pytorch_checkpoint_path", default=None, type=str, help=( "Path to the PyTorch checkpoint path or shortcut name to download from AWS. " "If not given, will download and convert all the checkpoints from AWS." ), ) parser.add_argument( "--config_file", default=None, type=str, help=( "The config json file corresponding to the pre-trained model. \n" "This specifies the model architecture. If not given and " "--pytorch_checkpoint_path is not given or is a shortcut name " "use the configuration associated to the shortcut name on the AWS" ), ) parser.add_argument( "--compare_with_pt_model", action="store_true", help="Compare Tensorflow and PyTorch model predictions." ) parser.add_argument( "--use_cached_models", action="store_true", help="Use cached models if possible instead of updating to latest checkpoint versions.", ) parser.add_argument( "--remove_cached_files", action="store_true", help="Remove pytorch models after conversion (save memory when converting in batches).", ) parser.add_argument("--only_convert_finetuned_models", action="store_true", help="Only convert finetuned models.") _UpperCAmelCase : Optional[int] = parser.parse_args() # if args.pytorch_checkpoint_path is not None: # convert_pt_checkpoint_to_tf(args.model_type.lower(), # args.pytorch_checkpoint_path, # args.config_file if args.config_file is not None else args.pytorch_checkpoint_path, # args.tf_dump_path, # compare_with_pt_model=args.compare_with_pt_model, # use_cached_models=args.use_cached_models) # else: convert_all_pt_checkpoints_to_tf( args.model_type.lower() if args.model_type is not None else None, args.tf_dump_path, model_shortcut_names_or_path=[args.pytorch_checkpoint_path] if args.pytorch_checkpoint_path is not None else None, config_shortcut_names_or_path=[args.config_file] if args.config_file is not None else None, compare_with_pt_model=args.compare_with_pt_model, use_cached_models=args.use_cached_models, remove_cached_files=args.remove_cached_files, only_convert_finetuned_models=args.only_convert_finetuned_models, )
363
import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class lowercase ( _SCREAMING_SNAKE_CASE , unittest.TestCase ): __lowercase : Any = CTRLTokenizer __lowercase : Any = False __lowercase : Union[str, Any] = False def __UpperCamelCase ( self ) -> Any: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase = ['adapt', 're@@', 'a@@', 'apt', 'c@@', 't', '<unk>'] UpperCamelCase = dict(zip(A_ , range(len(A_ ) ) ) ) UpperCamelCase = ['#version: 0.2', 'a p', 'ap t</w>', 'r e', 'a d', 'ad apt</w>', ''] UpperCamelCase = {'unk_token': '<unk>'} UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(A_ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(A_ ) ) def __UpperCamelCase ( self , **A_ ) -> List[Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **A_ ) def __UpperCamelCase ( self , A_ ) -> Optional[Any]: """simple docstring""" UpperCamelCase = 'adapt react readapt apt' UpperCamelCase = 'adapt react readapt apt' return input_text, output_text def __UpperCamelCase ( self ) -> str: """simple docstring""" UpperCamelCase = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) UpperCamelCase = 'adapt react readapt apt' UpperCamelCase = 'adapt re@@ a@@ c@@ t re@@ adapt apt'.split() UpperCamelCase = tokenizer.tokenize(A_ ) self.assertListEqual(A_ , A_ ) UpperCamelCase = tokens + [tokenizer.unk_token] UpperCamelCase = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(A_ ) , A_ )
110
0
'''simple docstring''' import math import sys def snake_case_ ( _lowerCAmelCase : str ) -> str: UpperCAmelCase : List[Any] = '''''' try: with open(_lowerCAmelCase , '''rb''' ) as binary_file: UpperCAmelCase : str = binary_file.read() for dat in data: UpperCAmelCase : Optional[int] = f"""{dat:08b}""" result += curr_byte return result except OSError: print('''File not accessible''' ) sys.exit() def snake_case_ ( _lowerCAmelCase : str ) -> str: UpperCAmelCase : Dict = {'''0''': '''0''', '''1''': '''1'''} UpperCAmelCase , UpperCAmelCase : Optional[Any] = '''''', '''''' UpperCAmelCase : int = len(_lowerCAmelCase ) for i in range(len(_lowerCAmelCase ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue UpperCAmelCase : List[Any] = lexicon[curr_string] result += last_match_id UpperCAmelCase : List[str] = last_match_id + '''0''' if math.loga(_lowerCAmelCase ).is_integer(): UpperCAmelCase : List[str] = {} for curr_key in list(_lowerCAmelCase ): UpperCAmelCase : Any = lexicon.pop(_lowerCAmelCase ) UpperCAmelCase : Dict = new_lex UpperCAmelCase : Optional[Any] = last_match_id + '''1''' index += 1 UpperCAmelCase : List[Any] = '''''' return result def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> None: UpperCAmelCase : Union[str, Any] = 8 try: with open(_lowerCAmelCase , '''wb''' ) as opened_file: UpperCAmelCase : Optional[int] = [ to_write[i : i + byte_length] for i in range(0 , len(_lowerCAmelCase ) , _lowerCAmelCase ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('''10000000''' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(_lowerCAmelCase , 2 ).to_bytes(1 , byteorder='''big''' ) ) except OSError: print('''File not accessible''' ) sys.exit() def snake_case_ ( _lowerCAmelCase : str ) -> str: UpperCAmelCase : List[Any] = 0 for letter in data_bits: if letter == "1": break counter += 1 UpperCAmelCase : Tuple = data_bits[counter:] UpperCAmelCase : Optional[Any] = data_bits[counter + 1 :] return data_bits def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> None: UpperCAmelCase : Optional[Any] = read_file_binary(_lowerCAmelCase ) UpperCAmelCase : Dict = remove_prefix(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = decompress_data(_lowerCAmelCase ) write_file_binary(_lowerCAmelCase , _lowerCAmelCase ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
23
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> bool: UpperCAmelCase : str = get_failure_array(_lowerCAmelCase ) # 2) Step through text searching for pattern UpperCAmelCase , UpperCAmelCase : Optional[Any] = 0, 0 # index into text, pattern while i < len(_lowerCAmelCase ): if pattern[j] == text[i]: if j == (len(_lowerCAmelCase ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: UpperCAmelCase : Optional[Any] = failure[j - 1] continue i += 1 return False def snake_case_ ( _lowerCAmelCase : str ) -> list[int]: UpperCAmelCase : Optional[Any] = [0] UpperCAmelCase : str = 0 UpperCAmelCase : List[str] = 1 while j < len(_lowerCAmelCase ): if pattern[i] == pattern[j]: i += 1 elif i > 0: UpperCAmelCase : Union[str, Any] = failure[i - 1] continue j += 1 failure.append(_lowerCAmelCase ) return failure if __name__ == "__main__": # Test 1) UpperCamelCase__: str = "abc1abc12" UpperCamelCase__: str = "alskfjaldsabc1abc1abc12k23adsfabcabc" UpperCamelCase__: Any = "alskfjaldsk23adsfabcabc" assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) UpperCamelCase__: Tuple = "ABABX" UpperCamelCase__: Union[str, Any] = "ABABZABABYABABX" assert kmp(pattern, text) # Test 3) UpperCamelCase__: Any = "AAAB" UpperCamelCase__: str = "ABAAAAAB" assert kmp(pattern, text) # Test 4) UpperCamelCase__: int = "abcdabcy" UpperCamelCase__: Any = "abcxabcdabxabcdabcdabcy" assert kmp(pattern, text) # Test 5) UpperCamelCase__: List[str] = "aabaabaaa" assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
23
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A : int = logging.get_logger(__name__) A : List[str] = { "facebook/timesformer": "https://huggingface.co/facebook/timesformer/resolve/main/config.json", } class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' __UpperCAmelCase : str ="""timesformer""" def __init__( self , __a=2_24 , __a=16 , __a=3 , __a=8 , __a=7_68 , __a=12 , __a=12 , __a=30_72 , __a="gelu" , __a=0.0 , __a=0.0 , __a=0.0_2 , __a=1e-6 , __a=True , __a="divided_space_time" , __a=0 , **__a , ): super().__init__(**__a ) __lowerCAmelCase = image_size __lowerCAmelCase = patch_size __lowerCAmelCase = num_channels __lowerCAmelCase = num_frames __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = qkv_bias __lowerCAmelCase = attention_type __lowerCAmelCase = drop_path_rate
259
"""simple docstring""" import string def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = "" for i in sequence: __lowerCAmelCase = ord(_UpperCamelCase ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extract <= 122: output += chr(219 - extract ) else: output += i return output def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = string.ascii_letters __lowerCAmelCase = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(_UpperCamelCase )] if c in letters else c for c in sequence ) def _lowerCamelCase ( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) __lowerCAmelCase = "from string import printable ; from __main__ import atbash, atbash_slow" print(f"> atbash_slow(): {timeit('atbash_slow(printable)' , setup=_UpperCamelCase )} seconds" ) print(f"> atbash(): {timeit('atbash(printable)' , setup=_UpperCamelCase )} seconds" ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(f'''{example} encrypted in atbash: {atbash(example)}''') benchmark()
259
1
import mpmath # for roots of unity import numpy as np class __snake_case : def __init__( self : str , _lowercase : List[Any]=None , _lowercase : List[Any]=None ): """simple docstring""" SCREAMING_SNAKE_CASE__ = list(poly_a or [0] )[:] SCREAMING_SNAKE_CASE__ = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() SCREAMING_SNAKE_CASE__ = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() SCREAMING_SNAKE_CASE__ = len(self.polyB ) # Add 0 to make lengths equal a power of 2 SCREAMING_SNAKE_CASE__ = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform SCREAMING_SNAKE_CASE__ = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product SCREAMING_SNAKE_CASE__ = self.__multiply() def __a ( self : Dict , _lowercase : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [[x] for x in self.polyA] if which == """A""" else [[x] for x in self.polyB] # Corner case if len(_lowercase ) <= 1: return dft[0] # SCREAMING_SNAKE_CASE__ = self.c_max_length // 2 while next_ncol > 0: SCREAMING_SNAKE_CASE__ = [[] for i in range(_lowercase )] SCREAMING_SNAKE_CASE__ = self.root**next_ncol # First half of next step SCREAMING_SNAKE_CASE__ = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(_lowercase ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step SCREAMING_SNAKE_CASE__ = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(_lowercase ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update SCREAMING_SNAKE_CASE__ = new_dft SCREAMING_SNAKE_CASE__ = next_ncol // 2 return dft[0] def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.__dft("""A""" ) SCREAMING_SNAKE_CASE__ = self.__dft("""B""" ) SCREAMING_SNAKE_CASE__ = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT SCREAMING_SNAKE_CASE__ = 2 while next_ncol <= self.c_max_length: SCREAMING_SNAKE_CASE__ = [[] for i in range(_lowercase )] SCREAMING_SNAKE_CASE__ = self.root ** (next_ncol // 2) SCREAMING_SNAKE_CASE__ = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update SCREAMING_SNAKE_CASE__ = new_inverse_c next_ncol *= 2 # Unpack SCREAMING_SNAKE_CASE__ = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1J for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """A = """ + """ + """.join( f"""{coef}*x^{i}""" for coef, i in enumerate(self.polyA[: self.len_A] ) ) SCREAMING_SNAKE_CASE__ = """B = """ + """ + """.join( f"""{coef}*x^{i}""" for coef, i in enumerate(self.polyB[: self.len_B] ) ) SCREAMING_SNAKE_CASE__ = """A*B = """ + """ + """.join( f"""{coef}*x^{i}""" for coef, i in enumerate(self.product ) ) return f"""{a}\n{b}\n{c}""" # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
219
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( lowerCamelCase_ , unittest.TestCase ): lowerCAmelCase_ = KandinskyVaaImgaImgPipeline lowerCAmelCase_ = ["image_embeds", "negative_image_embeds", "image"] lowerCAmelCase_ = [ "image_embeds", "negative_image_embeds", "image", ] lowerCAmelCase_ = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] lowerCAmelCase_ = False @property def __a ( self : Union[str, Any] ): """simple docstring""" return 32 @property def __a ( self : Union[str, Any] ): """simple docstring""" return 32 @property def __a ( self : Optional[Any] ): """simple docstring""" return self.time_input_dim @property def __a ( self : Optional[int] ): """simple docstring""" return self.time_input_dim * 4 @property def __a ( self : List[str] ): """simple docstring""" return 1_00 @property def __a ( self : Union[str, Any] ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = { """in_channels""": 4, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } SCREAMING_SNAKE_CASE__ = UNetaDConditionModel(**_lowercase ) return model @property def __a ( self : str ): """simple docstring""" return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __a ( self : Union[str, Any] ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = VQModel(**self.dummy_movq_kwargs ) return model def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.dummy_unet SCREAMING_SNAKE_CASE__ = self.dummy_movq SCREAMING_SNAKE_CASE__ = { """num_train_timesteps""": 10_00, """beta_schedule""": """linear""", """beta_start""": 0.0_00_85, """beta_end""": 0.0_12, """clip_sample""": False, """set_alpha_to_one""": False, """steps_offset""": 0, """prediction_type""": """epsilon""", """thresholding""": False, } SCREAMING_SNAKE_CASE__ = DDIMScheduler(**_lowercase ) SCREAMING_SNAKE_CASE__ = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def __a ( self : Optional[Any] , _lowercase : Any , _lowercase : Tuple=0 ): """simple docstring""" SCREAMING_SNAKE_CASE__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_lowercase ) ).to(_lowercase ) SCREAMING_SNAKE_CASE__ = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _lowercase ) # create init_image SCREAMING_SNAKE_CASE__ = floats_tensor((1, 3, 64, 64) , rng=random.Random(_lowercase ) ).to(_lowercase ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] SCREAMING_SNAKE_CASE__ = Image.fromarray(np.uinta(_lowercase ) ).convert("""RGB""" ).resize((2_56, 2_56) ) if str(_lowercase ).startswith("""mps""" ): SCREAMING_SNAKE_CASE__ = torch.manual_seed(_lowercase ) else: SCREAMING_SNAKE_CASE__ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) SCREAMING_SNAKE_CASE__ = { """image""": init_image, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 10, """guidance_scale""": 7.0, """strength""": 0.2, """output_type""": """np""", } return inputs def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """cpu""" SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = self.pipeline_class(**_lowercase ) SCREAMING_SNAKE_CASE__ = pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = pipe(**self.get_dummy_inputs(_lowercase ) ) SCREAMING_SNAKE_CASE__ = output.images SCREAMING_SNAKE_CASE__ = pipe( **self.get_dummy_inputs(_lowercase ) , return_dict=_lowercase , )[0] SCREAMING_SNAKE_CASE__ = image[0, -3:, -3:, -1] SCREAMING_SNAKE_CASE__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE__ = np.array( [0.6_19_97_78, 0.63_98_44_06, 0.46_14_57_85, 0.62_94_49_84, 0.5_62_22_15, 0.47_30_61_32, 0.47_44_14_56, 0.4_60_76_06, 0.48_71_92_63] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def __a ( self : Optional[int] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_img2img_frog.npy""" ) SCREAMING_SNAKE_CASE__ = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) SCREAMING_SNAKE_CASE__ = """A red cartoon frog, 4k""" SCREAMING_SNAKE_CASE__ = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(_lowercase ) SCREAMING_SNAKE_CASE__ = KandinskyVaaImgaImgPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder""" , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE__ = pipeline.to(_lowercase ) pipeline.set_progress_bar_config(disable=_lowercase ) SCREAMING_SNAKE_CASE__ = torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = pipe_prior( _lowercase , generator=_lowercase , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() SCREAMING_SNAKE_CASE__ = pipeline( image=_lowercase , image_embeds=_lowercase , negative_image_embeds=_lowercase , generator=_lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , strength=0.2 , output_type="""np""" , ) SCREAMING_SNAKE_CASE__ = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(_lowercase , _lowercase )
219
1
"""simple docstring""" import math import random def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase = False ) -> float: if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value __UpperCamelCase = 0.02 def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> float: snake_case_ = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(snake_case_ ): # Forward propagation snake_case_ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? snake_case_ = (expected / 100) - layer_a # Error delta snake_case_ = layer_1_error * sigmoid_function(snake_case_ , snake_case_ ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase = int(input('''Expected value: ''')) __UpperCamelCase = int(input('''Number of propagations: ''')) print(forward_propagation(expected, number_propagations))
368
"""simple docstring""" import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class UpperCamelCase ( lowerCAmelCase__ , unittest.TestCase ): SCREAMING_SNAKE_CASE_ = DebertaVaTokenizer SCREAMING_SNAKE_CASE_ = DebertaVaTokenizerFast SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = True def a_ ( self) -> int: super().setUp() # We have a SentencePiece fixture for testing snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, unk_token='<unk>') tokenizer.save_pretrained(self.tmpdirname) def a_ ( self, lowerCAmelCase__) -> Any: snake_case_ = 'this is a test' snake_case_ = 'this is a test' return input_text, output_text def a_ ( self) -> Optional[int]: snake_case_ = '<pad>' snake_case_ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__), lowerCAmelCase__) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__), lowerCAmelCase__) def a_ ( self) -> Tuple: snake_case_ = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0], '<pad>') self.assertEqual(vocab_keys[1], '<unk>') self.assertEqual(vocab_keys[-1], '[PAD]') self.assertEqual(len(lowerCAmelCase__), 3_0001) def a_ ( self) -> Dict: self.assertEqual(self.get_tokenizer().vocab_size, 3_0000) def a_ ( self) -> List[str]: # fmt: off snake_case_ = ' \tHeLLo!how \n Are yoU? ' snake_case_ = ['▁hello', '!', 'how', '▁are', '▁you', '?'] # fmt: on snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, do_lower_case=lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = DebertaVaTokenizerFast(lowerCAmelCase__, do_lower_case=lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.') def a_ ( self) -> str: pass @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.') def a_ ( self) -> List[Any]: pass def a_ ( self) -> str: # fmt: off snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = DebertaVaTokenizerFast(lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) def a_ ( self) -> List[Any]: # fmt: off snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = DebertaVaTokenizerFast(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) def a_ ( self) -> Dict: # fmt: off snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = DebertaVaTokenizerFast(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) def a_ ( self) -> Tuple: # fmt: off snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = DebertaVaTokenizerFast(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) def a_ ( self) -> Any: # fmt: off snake_case_ = ' \tHeLLo!how \n Are yoU? ' snake_case_ = ['▁', '<unk>', 'e', '<unk>', 'o', '!', 'how', '▁', '<unk>', 're', '▁yo', '<unk>', '?'] # fmt: on snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = DebertaVaTokenizerFast(lowerCAmelCase__, do_lower_case=lowerCAmelCase__, split_by_punct=lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) def a_ ( self) -> Dict: snake_case_ = self.get_tokenizer() snake_case_ = self.get_rust_tokenizer() snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = tokenizer.convert_ids_to_tokens(tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) snake_case_ = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__)) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__) snake_case_ = rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = self.get_rust_tokenizer() snake_case_ = tokenizer.encode(lowerCAmelCase__) snake_case_ = rust_tokenizer.encode(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) def a_ ( self) -> int: snake_case_ = 'This is a test' snake_case_ = [13, 1, 4398, 25, 21, 1289] snake_case_ = ['▁', 'T', 'his', '▁is', '▁a', '▁test'] snake_case_ = ['▁', '<unk>', 'his', '▁is', '▁a', '▁test'] snake_case_ = DebertaVaTokenizer(lowerCAmelCase__, keep_accents=lowerCAmelCase__) snake_case_ = DebertaVaTokenizerFast(lowerCAmelCase__, keep_accents=lowerCAmelCase__) snake_case_ = tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = tokenizer.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = rust_tokenizer.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) # fmt: off snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = [13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] snake_case_ = ['▁', 'I', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.', ] snake_case_ = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on snake_case_ = tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = tokenizer.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = tokenizer.convert_ids_to_tokens(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = rust_tokenizer.encode(lowerCAmelCase__, add_special_tokens=lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = rust_tokenizer.tokenize(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = rust_tokenizer.convert_ids_to_tokens(lowerCAmelCase__) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__) def a_ ( self) -> Tuple: snake_case_ = DebertaVaTokenizer(lowerCAmelCase__) snake_case_ = tokenizer.encode('sequence builders') snake_case_ = tokenizer.encode('multi-sequence build') snake_case_ = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__) snake_case_ = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__, lowerCAmelCase__) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id], lowerCAmelCase__) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id], lowerCAmelCase__, ) @slow def a_ ( self) -> Union[str, Any]: # fmt: off snake_case_ = {'input_ids': [[1, 3_9867, 36, 1_9390, 486, 27, 3_5052, 8_1436, 18, 6_0685, 1225, 7, 3_5052, 8_1436, 18, 9367, 1_6899, 18, 1_5937, 53, 594, 773, 18, 1_6287, 3_0465, 36, 1_5937, 6, 4_1139, 38, 3_6979, 6_0763, 191, 6, 3_4132, 99, 6, 5_0538, 390, 4_3230, 6, 3_4132, 2779, 2_0850, 14, 699, 1072, 1194, 36, 382, 1_0901, 53, 7, 699, 1072, 2084, 36, 2_0422, 630, 53, 19, 105, 3049, 1896, 1053, 1_6899, 1506, 11, 3_7978, 4243, 7, 1237, 3_1869, 200, 1_6566, 654, 6, 3_5052, 8_1436, 7, 5_5630, 1_3593, 4, 2], [1, 26, 1_5011, 13, 667, 8, 1053, 18, 2_3611, 1237, 7_2356, 1_2820, 34, 10_4134, 1209, 35, 1_3313, 6627, 21, 202, 347, 7, 164, 2399, 11, 46, 4485, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1232, 2864, 1_5785, 1_4951, 105, 5, 8581, 1250, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__, model_name='microsoft/deberta-v2-xlarge', revision='ad6e42c1532ddf3a15c39246b63f5559d558b670', )
312
0
"""simple docstring""" from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class _SCREAMING_SNAKE_CASE : UpperCAmelCase_ :List[str] UpperCAmelCase_ :Optional[str] = None # Automatically constructed UpperCAmelCase_ :ClassVar[str] = "dict" UpperCAmelCase_ :ClassVar[Any] = None UpperCAmelCase_ :str = field(default="Translation" , init=A__ , repr=A__ ) def __call__( self ) -> Any: return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def __lowerCAmelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value("""string""" ) for k in sorted(self.languages )} @dataclass class _SCREAMING_SNAKE_CASE : UpperCAmelCase_ :Optional[List] = None UpperCAmelCase_ :Optional[int] = None UpperCAmelCase_ :Optional[str] = None # Automatically constructed UpperCAmelCase_ :ClassVar[str] = "dict" UpperCAmelCase_ :ClassVar[Any] = None UpperCAmelCase_ :str = field(default="TranslationVariableLanguages" , init=A__ , repr=A__ ) def __lowerCAmelCase ( self ) -> Dict: lowerCAmelCase_ :List[str] = sorted(set(self.languages ) ) if self.languages else None lowerCAmelCase_ :str = len(self.languages ) if self.languages else None def __call__( self ) -> int: return pa.struct({"""language""": pa.list_(pa.string() ), """translation""": pa.list_(pa.string() )} ) def __lowerCAmelCase ( self , __A ) -> Any: lowerCAmelCase_ :str = set(self.languages ) if self.languages and set(__A ) - lang_set: raise ValueError( f"""Some languages in example ({", ".join(sorted(set(__A ) - lang_set ) )}) are not in valid set ({", ".join(__A )}).""" ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. lowerCAmelCase_ :List[Any] = [] for lang, text in translation_dict.items(): if isinstance(__A , __A ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. lowerCAmelCase_ , lowerCAmelCase_ :List[Any] = zip(*sorted(__A ) ) return {"language": languages, "translation": translations} def __lowerCAmelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value("""string""" ) ), "translation": Sequence(Value("""string""" ) ), }
84
import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel SCREAMING_SNAKE_CASE :Union[str, Any] = False SCREAMING_SNAKE_CASE :Any = True SCREAMING_SNAKE_CASE :Tuple = False if __name__ == "__main__": SCREAMING_SNAKE_CASE :Tuple = argparse.ArgumentParser() parser.add_argument( '--repo_path', default=None, type=str, required=True, help='The config json file corresponding to the architecture.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') SCREAMING_SNAKE_CASE :Union[str, Any] = parser.parse_args() SCREAMING_SNAKE_CASE :Dict = { 'image_size': 'sample_size', 'num_res_blocks': 'layers_per_block', 'block_channels': 'block_out_channels', 'down_blocks': 'down_block_types', 'up_blocks': 'up_block_types', 'downscale_freq_shift': 'freq_shift', 'resnet_num_groups': 'norm_num_groups', 'resnet_act_fn': 'act_fn', 'resnet_eps': 'norm_eps', 'num_head_channels': 'attention_head_dim', } SCREAMING_SNAKE_CASE :Optional[int] = { 'time_steps': 'time_proj', 'mid': 'mid_block', 'downsample_blocks': 'down_blocks', 'upsample_blocks': 'up_blocks', } SCREAMING_SNAKE_CASE :int = '' if has_file(args.repo_path, 'config.json') else 'unet' with open(os.path.join(args.repo_path, subfolder, 'config.json'), 'r', encoding='utf-8') as reader: SCREAMING_SNAKE_CASE :Dict = reader.read() SCREAMING_SNAKE_CASE :List[str] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, 'config.json'): SCREAMING_SNAKE_CASE :Optional[int] = UNetaDModel(**config) else: SCREAMING_SNAKE_CASE :Optional[Any] = UNetaDConditionModel if 'ldm-text2im-large-256' in args.repo_path else UNetaDModel SCREAMING_SNAKE_CASE :List[str] = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) SCREAMING_SNAKE_CASE :List[str] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: SCREAMING_SNAKE_CASE :Optional[Any] = config[key] del config[key] SCREAMING_SNAKE_CASE :Optional[Any] = [k.replace('UNetRes', '') for k in config['down_block_types']] SCREAMING_SNAKE_CASE :List[Any] = [k.replace('UNetRes', '') for k in config['up_block_types']] if do_only_weights: SCREAMING_SNAKE_CASE :Tuple = torch.load(os.path.join(args.repo_path, subfolder, 'diffusion_pytorch_model.bin')) SCREAMING_SNAKE_CASE :Any = {} for param_key, param_value in state_dict.items(): if param_key.endswith('.op.bias') or param_key.endswith('.op.weight'): continue SCREAMING_SNAKE_CASE :List[str] = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('.')[0] == key: SCREAMING_SNAKE_CASE :List[Any] = param_value SCREAMING_SNAKE_CASE :str = True if not has_changed: SCREAMING_SNAKE_CASE :List[str] = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
15
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available _lowerCamelCase : str = { '''configuration_ernie''': ['''ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ErnieConfig''', '''ErnieOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[int] = [ '''ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ErnieForCausalLM''', '''ErnieForMaskedLM''', '''ErnieForMultipleChoice''', '''ErnieForNextSentencePrediction''', '''ErnieForPreTraining''', '''ErnieForQuestionAnswering''', '''ErnieForSequenceClassification''', '''ErnieForTokenClassification''', '''ErnieModel''', '''ErniePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys _lowerCamelCase : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
356
def __lowerCamelCase (UpperCAmelCase__ : str , UpperCAmelCase__ : str = " " ): SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = 0 for index, char in enumerate(UpperCAmelCase__ ): if char == separator: split_words.append(string[last_index:index] ) SCREAMING_SNAKE_CASE = index + 1 elif index + 1 == len(UpperCAmelCase__ ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
206
0
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowerCAmelCase ( UpperCamelCase__ , unittest.TestCase ): """simple docstring""" snake_case_ = CodeGenTokenizer snake_case_ = CodeGenTokenizerFast snake_case_ = True snake_case_ = {'''add_prefix_space''': True} snake_case_ = False def lowercase_ ( self ) -> Tuple: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowerCamelCase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] __lowerCamelCase = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_ ) ) ) ) __lowerCamelCase = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __lowerCamelCase = {'unk_token': '<unk>'} __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(UpperCamelCase_ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(UpperCamelCase_ ) ) def lowercase_ ( self , **lowerCamelCase__ ) -> Optional[int]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def lowercase_ ( self , **lowerCamelCase__ ) -> Optional[int]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **UpperCamelCase_ ) def lowercase_ ( self , lowerCamelCase__ ) -> Tuple: '''simple docstring''' __lowerCamelCase = 'lower newer' __lowerCamelCase = 'lower newer' return input_text, output_text def lowercase_ ( self ) -> Dict: '''simple docstring''' __lowerCamelCase = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __lowerCamelCase = 'lower newer' __lowerCamelCase = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] __lowerCamelCase = tokenizer.tokenize(UpperCamelCase_ , add_prefix_space=UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) , UpperCamelCase_ ) def lowercase_ ( self ) -> List[Any]: '''simple docstring''' if not self.test_rust_tokenizer: return __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = self.get_rust_tokenizer(add_prefix_space=UpperCamelCase_ ) __lowerCamelCase = 'lower newer' # Testing tokenization __lowerCamelCase = tokenizer.tokenize(UpperCamelCase_ , add_prefix_space=UpperCamelCase_ ) __lowerCamelCase = rust_tokenizer.tokenize(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) # Testing conversion to ids without special tokens __lowerCamelCase = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ , add_prefix_space=UpperCamelCase_ ) __lowerCamelCase = rust_tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) # Testing conversion to ids with special tokens __lowerCamelCase = self.get_rust_tokenizer(add_prefix_space=UpperCamelCase_ ) __lowerCamelCase = tokenizer.encode(UpperCamelCase_ , add_prefix_space=UpperCamelCase_ ) __lowerCamelCase = rust_tokenizer.encode(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) # Testing the unknown token __lowerCamelCase = tokens + [rust_tokenizer.unk_token] __lowerCamelCase = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(UpperCamelCase_ ) , UpperCamelCase_ ) def lowercase_ ( self , *lowerCamelCase__ , **lowerCamelCase__ ) -> Any: '''simple docstring''' pass def lowercase_ ( self , lowerCamelCase__=15 ) -> Tuple: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCamelCase = self.rust_tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_ ) # Simple input __lowerCamelCase = 'This is a simple input' __lowerCamelCase = ['This is a simple input 1', 'This is a simple input 2'] __lowerCamelCase = ('This is a simple input', 'This is a pair') __lowerCamelCase = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(UpperCamelCase_ , tokenizer_r.encode , UpperCamelCase_ , max_length=UpperCamelCase_ , padding='max_length' ) # Simple input self.assertRaises(UpperCamelCase_ , tokenizer_r.encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding='max_length' ) # Simple input self.assertRaises( UpperCamelCase_ , tokenizer_r.batch_encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding='max_length' , ) # Pair input self.assertRaises(UpperCamelCase_ , tokenizer_r.encode , UpperCamelCase_ , max_length=UpperCamelCase_ , padding='max_length' ) # Pair input self.assertRaises(UpperCamelCase_ , tokenizer_r.encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding='max_length' ) # Pair input self.assertRaises( UpperCamelCase_ , tokenizer_r.batch_encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding='max_length' , ) def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>' ) # Simple input __lowerCamelCase = 'This is a simple input' __lowerCamelCase = ['This is a simple input looooooooong', 'This is a simple input'] __lowerCamelCase = ('This is a simple input', 'This is a pair') __lowerCamelCase = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] __lowerCamelCase = tokenizer.pad_token_id __lowerCamelCase = tokenizer(UpperCamelCase_ , padding='max_length' , max_length=30 , return_tensors='np' ) __lowerCamelCase = tokenizer(UpperCamelCase_ , padding=UpperCamelCase_ , truncate=UpperCamelCase_ , return_tensors='np' ) __lowerCamelCase = tokenizer(*UpperCamelCase_ , padding='max_length' , max_length=60 , return_tensors='np' ) __lowerCamelCase = tokenizer(UpperCamelCase_ , padding=UpperCamelCase_ , truncate=UpperCamelCase_ , return_tensors='np' ) # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['input_ids'] ) self.assertTrue(0 in out_s['attention_mask'] ) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0] ) self.assertFalse(0 in out_sa['attention_mask'][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1] ) self.assertTrue(0 in out_sa['attention_mask'][1] ) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['input_ids'] ) self.assertTrue(0 in out_p['attention_mask'] ) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0] ) self.assertFalse(0 in out_pa['attention_mask'][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1] ) self.assertTrue(0 in out_pa['attention_mask'][1] ) def lowercase_ ( self ) -> str: '''simple docstring''' __lowerCamelCase = '$$$' __lowerCamelCase = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=UpperCamelCase_ , add_bos_token=UpperCamelCase_ ) __lowerCamelCase = 'This is a simple input' __lowerCamelCase = ['This is a simple input 1', 'This is a simple input 2'] __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = tokenizer(UpperCamelCase_ ) __lowerCamelCase = tokenizer(UpperCamelCase_ ) self.assertEqual(out_s.input_ids[0] , UpperCamelCase_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) __lowerCamelCase = tokenizer.decode(out_s.input_ids ) __lowerCamelCase = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , UpperCamelCase_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def lowercase_ ( self ) -> str: '''simple docstring''' __lowerCamelCase = CodeGenTokenizer.from_pretrained('Salesforce/codegen-350M-mono' ) __lowerCamelCase = '\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#' __lowerCamelCase = '\nif len_a > len_b: result = a\nelse: result = b' __lowerCamelCase = tokenizer.encode(UpperCamelCase_ ) __lowerCamelCase = ['^#', re.escape('<|endoftext|>' ), '^\'\'\'', '^"""', '\n\n\n'] __lowerCamelCase = tokenizer.decode(UpperCamelCase_ , truncate_before_pattern=UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' pass
90
def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" return credit_card_number.startswith(('''34''', '''35''', '''37''', '''4''', '''5''', '''6''') ) def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = credit_card_number lowercase__ = 0 lowercase__ = len(SCREAMING_SNAKE_CASE ) - 2 for i in range(SCREAMING_SNAKE_CASE , -1 , -2 ): # double the value of every second digit lowercase__ = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 lowercase__ = cc_number[:i] + str(SCREAMING_SNAKE_CASE ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(SCREAMING_SNAKE_CASE ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = f'{credit_card_number} is an invalid credit card number because' if not credit_card_number.isdigit(): print(f'{error_message} it has nonnumerical characters.' ) return False if not 13 <= len(SCREAMING_SNAKE_CASE ) <= 16: print(f'{error_message} of its length.' ) return False if not validate_initial_digits(SCREAMING_SNAKE_CASE ): print(f'{error_message} of its first two digits.' ) return False if not luhn_validation(SCREAMING_SNAKE_CASE ): print(f'{error_message} it fails the Luhn check.' ) return False print(f'{credit_card_number} is a valid credit card number.' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number('4111111111111111') validate_credit_card_number('32323')
110
0
from __future__ import annotations _snake_case : Union[str, Any] = [] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): for i in range(len(__lowerCamelCase ) ): if board[row][i] == 1: return False for i in range(len(__lowerCamelCase ) ): if board[i][column] == 1: return False for i, j in zip(range(__lowerCamelCase , -1 , -1 ) , range(__lowerCamelCase , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(__lowerCamelCase , -1 , -1 ) , range(__lowerCamelCase , len(__lowerCamelCase ) ) ): if board[i][j] == 1: return False return True def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): if row >= len(__lowerCamelCase ): solution.append(__lowerCamelCase ) printboard(__lowerCamelCase ) print() return True for i in range(len(__lowerCamelCase ) ): if is_safe(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[Any] = 1 solve(__lowerCamelCase , row + 1 ) __snake_case : Union[str, Any] = 0 return False def lowerCAmelCase_ ( __lowerCamelCase ): for i in range(len(__lowerCamelCase ) ): for j in range(len(__lowerCamelCase ) ): if board[i][j] == 1: print("Q" , end=" " ) else: print("." , end=" " ) print() # n=int(input("The no. of queens")) _snake_case : List[str] = 8 _snake_case : Optional[int] = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print("The total no. of solutions are :", len(solution))
134
import argparse import os import torch from transformers.utils import WEIGHTS_NAME _snake_case : Union[str, Any] = ["small", "medium", "large"] _snake_case : List[Any] = "lm_head.decoder.weight" _snake_case : Optional[Any] = "lm_head.weight" def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = torch.load(__lowerCamelCase ) __snake_case : Dict = d.pop(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) torch.save(__lowerCamelCase , os.path.join(__lowerCamelCase , __lowerCamelCase ) ) if __name__ == "__main__": _snake_case : Dict = argparse.ArgumentParser() parser.add_argument("--dialogpt_path", default=".", type=str) _snake_case : Any = parser.parse_args() for MODEL in DIALOGPT_MODELS: _snake_case : Dict = os.path.join(args.dialogpt_path, f'''{MODEL}_ft.pkl''') _snake_case : List[str] = f'''./DialoGPT-{MODEL}''' convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
134
1
import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() __snake_case = logging.get_logger("""transformers.models.encodec""") __snake_case = { """quantizer.vq.layers.*._codebook.inited""": """quantizer.layers.*.codebook.inited""", """quantizer.vq.layers.*._codebook.cluster_size""": """quantizer.layers.*.codebook.cluster_size""", """quantizer.vq.layers.*._codebook.embed""": """quantizer.layers.*.codebook.embed""", """quantizer.vq.layers.*._codebook.embed_avg""": """quantizer.layers.*.codebook.embed_avg""", } __snake_case = { """encoder.model.0.conv.conv""": """encoder.layers.0.conv""", """encoder.model.1.block.1.conv.conv""": """encoder.layers.1.block.1.conv""", """encoder.model.1.block.3.conv.conv""": """encoder.layers.1.block.3.conv""", """encoder.model.1.shortcut.conv.conv""": """encoder.layers.1.shortcut.conv""", """encoder.model.3.conv.conv""": """encoder.layers.3.conv""", """encoder.model.4.block.1.conv.conv""": """encoder.layers.4.block.1.conv""", """encoder.model.4.block.3.conv.conv""": """encoder.layers.4.block.3.conv""", """encoder.model.4.shortcut.conv.conv""": """encoder.layers.4.shortcut.conv""", """encoder.model.6.conv.conv""": """encoder.layers.6.conv""", """encoder.model.7.block.1.conv.conv""": """encoder.layers.7.block.1.conv""", """encoder.model.7.block.3.conv.conv""": """encoder.layers.7.block.3.conv""", """encoder.model.7.shortcut.conv.conv""": """encoder.layers.7.shortcut.conv""", """encoder.model.9.conv.conv""": """encoder.layers.9.conv""", """encoder.model.10.block.1.conv.conv""": """encoder.layers.10.block.1.conv""", """encoder.model.10.block.3.conv.conv""": """encoder.layers.10.block.3.conv""", """encoder.model.10.shortcut.conv.conv""": """encoder.layers.10.shortcut.conv""", """encoder.model.12.conv.conv""": """encoder.layers.12.conv""", """encoder.model.13.lstm""": """encoder.layers.13.lstm""", """encoder.model.15.conv.conv""": """encoder.layers.15.conv""", } __snake_case = { """encoder.model.0.conv.norm""": """encoder.layers.0.norm""", """encoder.model.1.block.1.conv.norm""": """encoder.layers.1.block.1.norm""", """encoder.model.1.block.3.conv.norm""": """encoder.layers.1.block.3.norm""", """encoder.model.1.shortcut.conv.norm""": """encoder.layers.1.shortcut.norm""", """encoder.model.3.conv.norm""": """encoder.layers.3.norm""", """encoder.model.4.block.1.conv.norm""": """encoder.layers.4.block.1.norm""", """encoder.model.4.block.3.conv.norm""": """encoder.layers.4.block.3.norm""", """encoder.model.4.shortcut.conv.norm""": """encoder.layers.4.shortcut.norm""", """encoder.model.6.conv.norm""": """encoder.layers.6.norm""", """encoder.model.7.block.1.conv.norm""": """encoder.layers.7.block.1.norm""", """encoder.model.7.block.3.conv.norm""": """encoder.layers.7.block.3.norm""", """encoder.model.7.shortcut.conv.norm""": """encoder.layers.7.shortcut.norm""", """encoder.model.9.conv.norm""": """encoder.layers.9.norm""", """encoder.model.10.block.1.conv.norm""": """encoder.layers.10.block.1.norm""", """encoder.model.10.block.3.conv.norm""": """encoder.layers.10.block.3.norm""", """encoder.model.10.shortcut.conv.norm""": """encoder.layers.10.shortcut.norm""", """encoder.model.12.conv.norm""": """encoder.layers.12.norm""", """encoder.model.15.conv.norm""": """encoder.layers.15.norm""", } __snake_case = { """decoder.model.0.conv.conv""": """decoder.layers.0.conv""", """decoder.model.1.lstm""": """decoder.layers.1.lstm""", """decoder.model.3.convtr.convtr""": """decoder.layers.3.conv""", """decoder.model.4.block.1.conv.conv""": """decoder.layers.4.block.1.conv""", """decoder.model.4.block.3.conv.conv""": """decoder.layers.4.block.3.conv""", """decoder.model.4.shortcut.conv.conv""": """decoder.layers.4.shortcut.conv""", """decoder.model.6.convtr.convtr""": """decoder.layers.6.conv""", """decoder.model.7.block.1.conv.conv""": """decoder.layers.7.block.1.conv""", """decoder.model.7.block.3.conv.conv""": """decoder.layers.7.block.3.conv""", """decoder.model.7.shortcut.conv.conv""": """decoder.layers.7.shortcut.conv""", """decoder.model.9.convtr.convtr""": """decoder.layers.9.conv""", """decoder.model.10.block.1.conv.conv""": """decoder.layers.10.block.1.conv""", """decoder.model.10.block.3.conv.conv""": """decoder.layers.10.block.3.conv""", """decoder.model.10.shortcut.conv.conv""": """decoder.layers.10.shortcut.conv""", """decoder.model.12.convtr.convtr""": """decoder.layers.12.conv""", """decoder.model.13.block.1.conv.conv""": """decoder.layers.13.block.1.conv""", """decoder.model.13.block.3.conv.conv""": """decoder.layers.13.block.3.conv""", """decoder.model.13.shortcut.conv.conv""": """decoder.layers.13.shortcut.conv""", """decoder.model.15.conv.conv""": """decoder.layers.15.conv""", } __snake_case = { """decoder.model.0.conv.norm""": """decoder.layers.0.norm""", """decoder.model.3.convtr.norm""": """decoder.layers.3.norm""", """decoder.model.4.block.1.conv.norm""": """decoder.layers.4.block.1.norm""", """decoder.model.4.block.3.conv.norm""": """decoder.layers.4.block.3.norm""", """decoder.model.4.shortcut.conv.norm""": """decoder.layers.4.shortcut.norm""", """decoder.model.6.convtr.norm""": """decoder.layers.6.norm""", """decoder.model.7.block.1.conv.norm""": """decoder.layers.7.block.1.norm""", """decoder.model.7.block.3.conv.norm""": """decoder.layers.7.block.3.norm""", """decoder.model.7.shortcut.conv.norm""": """decoder.layers.7.shortcut.norm""", """decoder.model.9.convtr.norm""": """decoder.layers.9.norm""", """decoder.model.10.block.1.conv.norm""": """decoder.layers.10.block.1.norm""", """decoder.model.10.block.3.conv.norm""": """decoder.layers.10.block.3.norm""", """decoder.model.10.shortcut.conv.norm""": """decoder.layers.10.shortcut.norm""", """decoder.model.12.convtr.norm""": """decoder.layers.12.norm""", """decoder.model.13.block.1.conv.norm""": """decoder.layers.13.block.1.norm""", """decoder.model.13.block.3.conv.norm""": """decoder.layers.13.block.3.norm""", """decoder.model.13.shortcut.conv.norm""": """decoder.layers.13.shortcut.norm""", """decoder.model.15.conv.norm""": """decoder.layers.15.norm""", } __snake_case = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } __snake_case = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } __snake_case = [] __snake_case = [] def _A ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[Any] ): for attribute in key.split('''.''' ): UpperCamelCase :List[str] = getattr(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if weight_type is not None: UpperCamelCase :Union[str, Any] = getattr(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ).shape else: UpperCamelCase :Dict = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": UpperCamelCase :Any = value elif weight_type == "weight_g": UpperCamelCase :Any = value elif weight_type == "weight_v": UpperCamelCase :List[Any] = value elif weight_type == "bias": UpperCamelCase :int = value elif weight_type == "running_mean": UpperCamelCase :Dict = value elif weight_type == "running_var": UpperCamelCase :Optional[Any] = value elif weight_type == "num_batches_tracked": UpperCamelCase :int = value elif weight_type == "weight_ih_l0": UpperCamelCase :List[str] = value elif weight_type == "weight_hh_l0": UpperCamelCase :Union[str, Any] = value elif weight_type == "bias_ih_l0": UpperCamelCase :int = value elif weight_type == "bias_hh_l0": UpperCamelCase :List[Any] = value elif weight_type == "weight_ih_l1": UpperCamelCase :Optional[Any] = value elif weight_type == "weight_hh_l1": UpperCamelCase :Dict = value elif weight_type == "bias_ih_l1": UpperCamelCase :Optional[int] = value elif weight_type == "bias_hh_l1": UpperCamelCase :Any = value else: UpperCamelCase :Optional[Any] = value logger.info(F'''{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.''' ) def _A ( SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Union[str, Any] ): for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: UpperCamelCase , UpperCamelCase :Dict = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def _A ( SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Any ): UpperCamelCase :Union[str, Any] = [] if model_name == "encodec_24khz" or "encodec_32khz": UpperCamelCase :List[Any] = MAPPING_24K elif model_name == "encodec_48khz": UpperCamelCase :Optional[Any] = MAPPING_48K else: raise ValueError(F'''Unsupported model: {model_name}''' ) for name, value in orig_dict.items(): if should_ignore(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): logger.info(F'''{name} was ignored''' ) continue UpperCamelCase :List[str] = False for key, mapped_key in MAPPING.items(): if "*" in key: UpperCamelCase , UpperCamelCase :Dict = key.split('''.*.''' ) if prefix in name and suffix in name: UpperCamelCase :str = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('''embed''' ) and name.endswith('''embed_avg''' ): continue UpperCamelCase :Dict = True if "*" in mapped_key: UpperCamelCase :Any = name.split(SCREAMING_SNAKE_CASE__ )[0].split('''.''' )[-2] UpperCamelCase :List[str] = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE__ ) if "weight_g" in name: UpperCamelCase :Tuple = '''weight_g''' elif "weight_v" in name: UpperCamelCase :Tuple = '''weight_v''' elif "weight_ih_l0" in name: UpperCamelCase :Any = '''weight_ih_l0''' elif "weight_hh_l0" in name: UpperCamelCase :Optional[int] = '''weight_hh_l0''' elif "bias_ih_l0" in name: UpperCamelCase :Optional[int] = '''bias_ih_l0''' elif "bias_hh_l0" in name: UpperCamelCase :str = '''bias_hh_l0''' elif "weight_ih_l1" in name: UpperCamelCase :Optional[Any] = '''weight_ih_l1''' elif "weight_hh_l1" in name: UpperCamelCase :Optional[int] = '''weight_hh_l1''' elif "bias_ih_l1" in name: UpperCamelCase :Optional[Any] = '''bias_ih_l1''' elif "bias_hh_l1" in name: UpperCamelCase :Any = '''bias_hh_l1''' elif "bias" in name: UpperCamelCase :Union[str, Any] = '''bias''' elif "weight" in name: UpperCamelCase :Union[str, Any] = '''weight''' elif "running_mean" in name: UpperCamelCase :Tuple = '''running_mean''' elif "running_var" in name: UpperCamelCase :Optional[Any] = '''running_var''' elif "num_batches_tracked" in name: UpperCamelCase :Optional[int] = '''num_batches_tracked''' else: UpperCamelCase :List[str] = None set_recursively(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE__ ) logger.warning(F'''Unused weights: {unused_weights}''' ) @torch.no_grad() def _A ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : str=None , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , ): if config_path is not None: UpperCamelCase :List[str] = EncodecConfig.from_pretrained(SCREAMING_SNAKE_CASE__ ) else: UpperCamelCase :List[Any] = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": UpperCamelCase :List[str] = [8, 5, 4, 4] UpperCamelCase :Any = [2.2] UpperCamelCase :Any = 64 UpperCamelCase :Optional[int] = 32000 UpperCamelCase :Union[str, Any] = 2048 UpperCamelCase :Any = False UpperCamelCase :List[Any] = False UpperCamelCase :List[str] = False elif model_name == "encodec_48khz": UpperCamelCase :Tuple = [8, 5, 4, 2] UpperCamelCase :Union[str, Any] = [3.0, 6.0, 12.0, 24.0] UpperCamelCase :Any = 48000 UpperCamelCase :List[Any] = 2 UpperCamelCase :str = False UpperCamelCase :List[Any] = '''time_group_norm''' UpperCamelCase :Union[str, Any] = True UpperCamelCase :Any = 1.0 UpperCamelCase :Union[str, Any] = 0.01 else: raise ValueError(F'''Unknown model name: {model_name}''' ) UpperCamelCase :Optional[Any] = EncodecModel(SCREAMING_SNAKE_CASE__ ) UpperCamelCase :Optional[Any] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(SCREAMING_SNAKE_CASE__ ) UpperCamelCase :List[str] = torch.load(SCREAMING_SNAKE_CASE__ ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights UpperCamelCase :List[Any] = original_checkpoint['''best_state'''] recursively_load_weights(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) if repo_id: print('''Pushing to the hub...''' ) feature_extractor.push_to_hub(SCREAMING_SNAKE_CASE__ ) model.push_to_hub(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument( """--model""", default="""encodec_24khz""", type=str, help="""The model to convert. Should be one of 'encodec_24khz', 'encodec_32khz', 'encodec_48khz'.""", ) parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) __snake_case = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
259
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class UpperCAmelCase_ ( lowercase ): """simple docstring""" UpperCamelCase_ : Dict ='git_vision_model' def __init__( self , SCREAMING_SNAKE_CASE_=768 , SCREAMING_SNAKE_CASE_=3072 , SCREAMING_SNAKE_CASE_=12 , SCREAMING_SNAKE_CASE_=12 , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=224 , SCREAMING_SNAKE_CASE_=16 , SCREAMING_SNAKE_CASE_="quick_gelu" , SCREAMING_SNAKE_CASE_=1e-5 , SCREAMING_SNAKE_CASE_=0.0 , SCREAMING_SNAKE_CASE_=0.02 , **SCREAMING_SNAKE_CASE_ , ) -> Tuple: super().__init__(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Optional[int] = hidden_size UpperCamelCase :Union[str, Any] = intermediate_size UpperCamelCase :Dict = num_hidden_layers UpperCamelCase :int = num_attention_heads UpperCamelCase :List[str] = num_channels UpperCamelCase :Optional[int] = patch_size UpperCamelCase :Optional[int] = image_size UpperCamelCase :List[Any] = initializer_range UpperCamelCase :Union[str, Any] = attention_dropout UpperCamelCase :Tuple = layer_norm_eps UpperCamelCase :Optional[Any] = hidden_act @classmethod def UpperCAmelCase ( cls , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) -> "PretrainedConfig": cls._set_token_in_kwargs(SCREAMING_SNAKE_CASE_ ) UpperCamelCase , UpperCamelCase :Dict = cls.get_config_dict(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) # get the vision config dict if we are loading from GITConfig if config_dict.get('''model_type''' ) == "git": UpperCamelCase :Tuple = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) class UpperCAmelCase_ ( lowercase ): """simple docstring""" UpperCamelCase_ : Optional[Any] ='git' def __init__( self , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=3_0522 , SCREAMING_SNAKE_CASE_=768 , SCREAMING_SNAKE_CASE_=6 , SCREAMING_SNAKE_CASE_=12 , SCREAMING_SNAKE_CASE_=3072 , SCREAMING_SNAKE_CASE_="gelu" , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=1024 , SCREAMING_SNAKE_CASE_=0.02 , SCREAMING_SNAKE_CASE_=1e-12 , SCREAMING_SNAKE_CASE_=0 , SCREAMING_SNAKE_CASE_="absolute" , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=101 , SCREAMING_SNAKE_CASE_=102 , SCREAMING_SNAKE_CASE_=None , **SCREAMING_SNAKE_CASE_ , ) -> int: super().__init__(bos_token_id=SCREAMING_SNAKE_CASE_ , eos_token_id=SCREAMING_SNAKE_CASE_ , pad_token_id=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) if vision_config is None: UpperCamelCase :Tuple = {} logger.info('''vision_config is None. initializing the GitVisionConfig with default values.''' ) UpperCamelCase :Union[str, Any] = GitVisionConfig(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Optional[int] = vocab_size UpperCamelCase :Optional[Any] = hidden_size UpperCamelCase :List[Any] = num_hidden_layers UpperCamelCase :List[Any] = num_attention_heads UpperCamelCase :Dict = hidden_act UpperCamelCase :List[str] = intermediate_size UpperCamelCase :List[str] = hidden_dropout_prob UpperCamelCase :Optional[int] = attention_probs_dropout_prob UpperCamelCase :Optional[Any] = max_position_embeddings UpperCamelCase :Tuple = initializer_range UpperCamelCase :Any = layer_norm_eps UpperCamelCase :int = position_embedding_type UpperCamelCase :Dict = use_cache UpperCamelCase :Tuple = tie_word_embeddings UpperCamelCase :Union[str, Any] = num_image_with_embedding UpperCamelCase :Optional[int] = bos_token_id UpperCamelCase :List[Any] = eos_token_id def UpperCAmelCase ( self ) -> Optional[int]: UpperCamelCase :Union[str, Any] = copy.deepcopy(self.__dict__ ) UpperCamelCase :Optional[int] = self.vision_config.to_dict() UpperCamelCase :int = self.__class__.model_type return output
259
1
SCREAMING_SNAKE_CASE_:Optional[Any] = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ SCREAMING_SNAKE_CASE_:Optional[Any] = [{"""type""": """code""", """content""": INSTALL_CONTENT}] SCREAMING_SNAKE_CASE_:List[str] = { """{processor_class}""": """FakeProcessorClass""", """{model_class}""": """FakeModelClass""", """{object_class}""": """FakeObjectClass""", }
115
import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self, lowerCamelCase__, lowerCamelCase__=14, lowerCamelCase__=7, lowerCamelCase__=True, lowerCamelCase__=True, lowerCamelCase__=False, lowerCamelCase__=True, lowerCamelCase__=99, lowerCamelCase__=32, lowerCamelCase__=4, lowerCamelCase__=4, lowerCamelCase__=4, lowerCamelCase__=37, lowerCamelCase__="gelu", lowerCamelCase__=0.1, lowerCamelCase__=0.1, lowerCamelCase__=512, lowerCamelCase__=0.02, ): A : List[str] = parent A : Any = batch_size A : Dict = seq_length A : Tuple = is_training A : Any = use_input_mask A : Any = use_token_type_ids A : Any = use_labels A : Optional[int] = vocab_size A : Dict = hidden_size A : Dict = rotary_dim A : Dict = num_hidden_layers A : Tuple = num_attention_heads A : Tuple = intermediate_size A : Union[str, Any] = hidden_act A : Dict = hidden_dropout_prob A : List[str] = attention_probs_dropout_prob A : Optional[int] = max_position_embeddings A : str = initializer_range A : Any = None A : Any = vocab_size - 1 A : int = vocab_size - 1 A : int = vocab_size - 1 def _lowerCAmelCase ( self ): A : str = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) A : Optional[int] = None if self.use_input_mask: A : Any = random_attention_mask([self.batch_size, self.seq_length] ) A : int = GPTJConfig( vocab_size=self.vocab_size, n_embd=self.hidden_size, n_layer=self.num_hidden_layers, n_head=self.num_attention_heads, n_positions=self.max_position_embeddings, use_cache=lowerCamelCase__, bos_token_id=self.bos_token_id, eos_token_id=self.eos_token_id, pad_token_id=self.pad_token_id, rotary_dim=self.rotary_dim, ) return (config, input_ids, input_mask) def _lowerCAmelCase ( self ): A : List[str] = self.prepare_config_and_inputs() A , A , A : List[str] = config_and_inputs A : Optional[int] = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ): A : Optional[int] = 20 A : Tuple = model_class_name(lowerCamelCase__ ) A : Dict = model.init_cache(input_ids.shape[0], lowerCamelCase__ ) A : int = jnp.ones((input_ids.shape[0], max_decoder_length), dtype="""i4""" ) A : Optional[int] = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :], (input_ids.shape[0], input_ids.shape[-1] - 1) ) A : List[Any] = model( input_ids[:, :-1], attention_mask=lowerCamelCase__, past_key_values=lowerCamelCase__, position_ids=lowerCamelCase__, ) A : List[Any] = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]], dtype="""i4""" ) A : Any = model( input_ids[:, -1:], attention_mask=lowerCamelCase__, past_key_values=outputs_cache.past_key_values, position_ids=lowerCamelCase__, ) A : Any = model(lowerCamelCase__ ) A : List[Any] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3, msg=f'''Max diff is {diff}''' ) def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ): A : Any = 20 A : Any = model_class_name(lowerCamelCase__ ) A : Dict = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )], axis=-1, ) A : str = model.init_cache(input_ids.shape[0], lowerCamelCase__ ) A : Any = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :], (input_ids.shape[0], input_ids.shape[-1] - 1) ) A : Optional[int] = model( input_ids[:, :-1], attention_mask=lowerCamelCase__, past_key_values=lowerCamelCase__, position_ids=lowerCamelCase__, ) A : str = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]], dtype="""i4""" ) A : List[Any] = model( input_ids[:, -1:], past_key_values=outputs_cache.past_key_values, attention_mask=lowerCamelCase__, position_ids=lowerCamelCase__, ) A : Union[str, Any] = model(lowerCamelCase__, attention_mask=lowerCamelCase__ ) A : str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3, msg=f'''Max diff is {diff}''' ) @require_flax class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __lowerCamelCase : Any = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () __lowerCamelCase : Optional[int] = (FlaxGPTJForCausalLM,) if is_flax_available() else () def _lowerCAmelCase ( self ): A : List[Any] = FlaxGPTJModelTester(self ) def _lowerCAmelCase ( self ): for model_class_name in self.all_model_classes: A , A , A : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) def _lowerCAmelCase ( self ): for model_class_name in self.all_model_classes: A , A , A : str = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) @tooslow def _lowerCAmelCase ( self ): A : int = GPTaTokenizer.from_pretrained("""gpt2""", pad_token="""<|endoftext|>""", padding_side="""left""" ) A : Optional[int] = tokenizer(["""Hello this is a long string""", """Hey"""], return_tensors="""np""", padding=lowerCamelCase__, truncation=lowerCamelCase__ ) A : Dict = FlaxGPTJForCausalLM.from_pretrained("""EleutherAI/gpt-j-6B""" ) A : str = False A : Optional[Any] = model.config.eos_token_id A : Union[str, Any] = jax.jit(model.generate ) A : str = jit_generate( inputs["""input_ids"""], attention_mask=inputs["""attention_mask"""], pad_token_id=tokenizer.pad_token_id ).sequences A : Optional[Any] = tokenizer.batch_decode(lowerCamelCase__, skip_special_tokens=lowerCamelCase__ ) A : Tuple = [ """Hello this is a long string of text.\n\nI'm trying to get the text of the""", """Hey, I'm a little late to the party. I'm going to""", ] self.assertListEqual(lowerCamelCase__, lowerCamelCase__ ) @is_pt_flax_cross_test def _lowerCAmelCase ( self ): A , A : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs A : Any = self._prepare_for_class(lowerCamelCase__, lowerCamelCase__ ) A : Dict = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class A : List[Any] = model_class.__name__[4:] # Skip the "Flax" at the beginning A : str = getattr(lowerCamelCase__, lowerCamelCase__ ) A , A : Optional[int] = pt_inputs["""input_ids"""].shape A : List[str] = np.random.randint(0, seq_length - 1, size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCamelCase__ ): A : List[Any] = 0 A : Tuple = 1 A : Optional[int] = 0 A : str = 1 A : Dict = pt_model_class(lowerCamelCase__ ).eval() A : int = model_class(lowerCamelCase__, dtype=jnp.floataa ) A : int = convert_pytorch_state_dict_to_flax(pt_model.state_dict(), lowerCamelCase__ ) A : Dict = fx_state with torch.no_grad(): A : Optional[int] = pt_model(**lowerCamelCase__ ).to_tuple() A : str = fx_model(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ), len(lowerCamelCase__ ), """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(lowerCamelCase__, lowerCamelCase__ ): self.assert_almost_equals(fx_output[:, -1], pt_output[:, -1].numpy(), 4e-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(lowerCamelCase__ ) A : Union[str, Any] = model_class.from_pretrained(lowerCamelCase__, from_pt=lowerCamelCase__ ) A : Any = fx_model_loaded(**lowerCamelCase__ ).to_tuple() self.assertEqual( len(lowerCamelCase__ ), len(lowerCamelCase__ ), """Output lengths differ between Flax and PyTorch""" ) for fx_output_loaded, pt_output in zip(lowerCamelCase__, lowerCamelCase__ ): self.assert_almost_equals(fx_output_loaded[:, -1], pt_output[:, -1].numpy(), 4e-2 ) @is_pt_flax_cross_test def _lowerCAmelCase ( self ): A , A : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs A : int = self._prepare_for_class(lowerCamelCase__, lowerCamelCase__ ) A : List[str] = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class A : Dict = model_class.__name__[4:] # Skip the "Flax" at the beginning A : Dict = getattr(lowerCamelCase__, lowerCamelCase__ ) A : int = pt_model_class(lowerCamelCase__ ).eval() A : int = model_class(lowerCamelCase__, dtype=jnp.floataa ) A : List[str] = load_flax_weights_in_pytorch_model(lowerCamelCase__, fx_model.params ) A , A : Optional[int] = pt_inputs["""input_ids"""].shape A : Optional[int] = np.random.randint(0, seq_length - 1, size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCamelCase__ ): A : Tuple = 0 A : Tuple = 1 A : str = 0 A : int = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): A : List[str] = pt_model(**lowerCamelCase__ ).to_tuple() A : Optional[int] = fx_model(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ), len(lowerCamelCase__ ), """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(lowerCamelCase__, lowerCamelCase__ ): self.assert_almost_equals(fx_output[:, -1], pt_output[:, -1].numpy(), 4e-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(lowerCamelCase__ ) A : str = pt_model_class.from_pretrained(lowerCamelCase__, from_flax=lowerCamelCase__ ) with torch.no_grad(): A : str = pt_model_loaded(**lowerCamelCase__ ).to_tuple() self.assertEqual( len(lowerCamelCase__ ), len(lowerCamelCase__ ), """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(lowerCamelCase__, lowerCamelCase__ ): self.assert_almost_equals(fx_output[:, -1], pt_output[:, -1].numpy(), 4e-2 ) @tooslow def _lowerCAmelCase ( self ): for model_class_name in self.all_model_classes: A : Union[str, Any] = model_class_name.from_pretrained("""EleutherAI/gpt-j-6B""" ) A : List[str] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ )
115
1
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( A_ ): # if the collection is empty, returns empty if collection == []: return [] # get some information about the collection lowerCAmelCase__ : Optional[Any] = len(A_ ) lowerCAmelCase__ : Tuple = max(A_ ) lowerCAmelCase__ : Optional[int] = min(A_ ) # create the counting array lowerCAmelCase__ : Optional[int] = coll_max + 1 - coll_min lowerCAmelCase__ : List[Any] = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , A_ ): lowerCAmelCase__ : Dict = counting_arr[i] + counting_arr[i - 1] # create the output collection lowerCAmelCase__ : Dict = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , A_ ) ): lowerCAmelCase__ : List[str] = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def __SCREAMING_SNAKE_CASE ( A_ ): return "".join([chr(A_ ) for i in counting_sort([ord(A_ ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string('''thisisthestring''') == "eghhiiinrsssttt" __UpperCamelCase : Optional[int] = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase : Dict = [int(item) for item in user_input.split(''',''')] print(counting_sort(unsorted))
106
import os from typing import Dict, List, Tuple, TypeVar, Union __a :Any = TypeVar('T') __a :Union[str, Any] = Union[List[T], Tuple[T, ...]] __a :List[str] = Union[T, List[T], Dict[str, T]] __a :Any = Union[str, bytes, os.PathLike]
312
0
from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class __lowerCAmelCase ( A ): UpperCamelCase = '''new-model''' if is_tf_available(): class __lowerCAmelCase ( A ): UpperCamelCase = NewModelConfig @require_tf class __lowerCAmelCase ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[Any]) -> Any: """simple docstring""" _UpperCAmelCase = 'bert-base-cased' _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModel.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow def _lowerCamelCase ( self : Tuple) -> Dict: """simple docstring""" _UpperCAmelCase = 'bert-base-cased' _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelForPreTraining.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow def _lowerCamelCase ( self : Any) -> Any: """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelForCausalLM.from_pretrained(A) _UpperCAmelCase , _UpperCAmelCase = TFAutoModelForCausalLM.from_pretrained(A , output_loading_info=A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow def _lowerCamelCase ( self : List[str]) -> Tuple: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelWithLMHead.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow def _lowerCamelCase ( self : Optional[int]) -> Any: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelForMaskedLM.from_pretrained(A) _UpperCAmelCase , _UpperCAmelCase = TFAutoModelForMaskedLM.from_pretrained(A , output_loading_info=A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow def _lowerCamelCase ( self : Optional[int]) -> Optional[int]: """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(A) _UpperCAmelCase , _UpperCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(A , output_loading_info=A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow def _lowerCamelCase ( self : Optional[Any]) -> int: """simple docstring""" for model_name in ["bert-base-uncased"]: _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelForSequenceClassification.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow def _lowerCamelCase ( self : Dict) -> Dict: """simple docstring""" for model_name in ["bert-base-uncased"]: _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelForQuestionAnswering.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) @slow @require_tensorflow_probability def _lowerCamelCase ( self : List[str]) -> Dict: """simple docstring""" for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _UpperCAmelCase = AutoConfig.from_pretrained(A) self.assertIsNotNone(A) self.assertIsInstance(A , A) _UpperCAmelCase = TFAutoModelForTableQuestionAnswering.from_pretrained(A) _UpperCAmelCase , _UpperCAmelCase = TFAutoModelForTableQuestionAnswering.from_pretrained( A , output_loading_info=A) self.assertIsNotNone(A) self.assertIsInstance(A , A) def _lowerCamelCase ( self : Optional[int]) -> Tuple: """simple docstring""" _UpperCAmelCase = TFAutoModelWithLMHead.from_pretrained(A) self.assertIsInstance(A , A) self.assertEqual(model.num_parameters() , 1_44_10) self.assertEqual(model.num_parameters(only_trainable=A) , 1_44_10) def _lowerCamelCase ( self : List[str]) -> Any: """simple docstring""" _UpperCAmelCase = TFAutoModelWithLMHead.from_pretrained(A) self.assertIsInstance(A , A) self.assertEqual(model.num_parameters() , 1_44_10) self.assertEqual(model.num_parameters(only_trainable=A) , 1_44_10) def _lowerCamelCase ( self : int) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = TFAutoModel.from_pretrained('sgugger/funnel-random-tiny') self.assertIsInstance(A , A) _UpperCAmelCase = copy.deepcopy(model.config) _UpperCAmelCase = ['FunnelBaseModel'] _UpperCAmelCase = TFAutoModel.from_config(A) self.assertIsInstance(A , A) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(A) _UpperCAmelCase = TFAutoModel.from_pretrained(A) self.assertIsInstance(A , A) def _lowerCamelCase ( self : Any) -> Any: """simple docstring""" try: AutoConfig.register('new-model' , A) _UpperCAmelCase = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__): # Wrong config class will raise an error with self.assertRaises(A): auto_class.register(A , A) auto_class.register(A , A) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(A): auto_class.register(A , A) # Now that the config is registered, it can be used as any other config with the auto-API _UpperCAmelCase = BertModelTester(self).get_config() _UpperCAmelCase = NewModelConfig(**tiny_config.to_dict()) _UpperCAmelCase = auto_class.from_config(A) self.assertIsInstance(A , A) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(A) _UpperCAmelCase = auto_class.from_pretrained(A) self.assertIsInstance(A , A) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def _lowerCamelCase ( self : Tuple) -> Any: """simple docstring""" with self.assertRaisesRegex( A , 'bert-base is not a local folder and is not a valid model identifier'): _UpperCAmelCase = TFAutoModel.from_pretrained('bert-base') def _lowerCamelCase ( self : Union[str, Any]) -> Any: """simple docstring""" with self.assertRaisesRegex( A , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)'): _UpperCAmelCase = TFAutoModel.from_pretrained(A , revision='aaaaaa') def _lowerCamelCase ( self : List[Any]) -> Optional[Any]: """simple docstring""" with self.assertRaisesRegex( A , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): _UpperCAmelCase = TFAutoModel.from_pretrained('hf-internal-testing/config-no-model') def _lowerCamelCase ( self : Any) -> Optional[Any]: """simple docstring""" with self.assertRaisesRegex(A , 'Use `from_pt=True` to load this model'): _UpperCAmelCase = TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only') def _lowerCamelCase ( self : Any) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert') with RequestCounter() as counter: _UpperCAmelCase = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert') self.assertEqual(counter.get_request_count , 0) self.assertEqual(counter.head_request_count , 1) self.assertEqual(counter.other_request_count , 0) # With a sharded checkpoint _UpperCAmelCase = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded') with RequestCounter() as counter: _UpperCAmelCase = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded') self.assertEqual(counter.get_request_count , 0) self.assertEqual(counter.head_request_count , 1) self.assertEqual(counter.other_request_count , 0)
290
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class __lowerCAmelCase : def __init__( self : Any , A : str = "cpu" , A : str = "openai/clip-vit-large-patch14") -> None: """simple docstring""" _UpperCAmelCase = device _UpperCAmelCase = CLIPTokenizerFast.from_pretrained(A) _UpperCAmelCase = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] _UpperCAmelCase = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] _UpperCAmelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std) _UpperCAmelCase = torchvision.transforms.Resize(2_24) _UpperCAmelCase = torchvision.transforms.CenterCrop(2_24) def _lowerCamelCase ( self : str , A : Any) -> str: """simple docstring""" _UpperCAmelCase = self.resize(A) _UpperCAmelCase = self.center_crop(A) _UpperCAmelCase = self.normalize(A) return images def __call__( self : Any , A : Dict=None , A : Dict=None , **A : List[Any]) -> Dict: """simple docstring""" _UpperCAmelCase = self.tokenizer(text=A , **A) _UpperCAmelCase = self.preprocess_img(A) _UpperCAmelCase = {key: value.to(self.device) for (key, value) in encoding.items()} return encoding class __lowerCAmelCase ( nn.Module ): def __init__( self : List[Any] , A : Any=10 , A : List[Any]=0.0_1 , A : Optional[int]=None , A : int=None , A : Dict=None , A : Tuple=None , A : str=None , A : Dict=None , A : Union[str, Any]=False , A : Any=True , A : Any="image" , A : Tuple=True , A : List[Any]=False , A : int=False , A : int=False , ) -> None: """simple docstring""" super().__init__() _UpperCAmelCase = None _UpperCAmelCase = device if device else get_device() if vqgan: _UpperCAmelCase = vqgan else: _UpperCAmelCase = load_vqgan(self.device , conf_path=A , ckpt_path=A) self.vqgan.eval() if clip: _UpperCAmelCase = clip else: _UpperCAmelCase = CLIPModel.from_pretrained('openai/clip-vit-base-patch32') self.clip.to(self.device) _UpperCAmelCase = ProcessorGradientFlow(device=self.device) _UpperCAmelCase = iterations _UpperCAmelCase = lr _UpperCAmelCase = log _UpperCAmelCase = make_grid _UpperCAmelCase = return_val _UpperCAmelCase = quantize _UpperCAmelCase = self.vqgan.decoder.z_shape def _lowerCamelCase ( self : Optional[int] , A : int=None , A : Union[str, Any]=None , A : Dict=5 , A : Optional[Any]=True) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = [] if output_path is None: _UpperCAmelCase = './animation.gif' if input_path is None: _UpperCAmelCase = self.save_path _UpperCAmelCase = sorted(glob(input_path + '/*')) if not len(A): raise ValueError( 'No images found in save path, aborting (did you pass save_intermediate=True to the generate' ' function?)') if len(A) == 1: print('Only one image found in save path, (did you pass save_intermediate=True to the generate function?)') _UpperCAmelCase = total_duration / len(A) _UpperCAmelCase = [frame_duration] * len(A) if extend_frames: _UpperCAmelCase = 1.5 _UpperCAmelCase = 3 for file_name in paths: if file_name.endswith('.png'): images.append(imageio.imread(A)) imageio.mimsave(A , A , duration=A) print(F"gif saved to {output_path}") def _lowerCamelCase ( self : List[str] , A : Optional[Any]=None , A : Optional[int]=None) -> int: """simple docstring""" if not (path or img): raise ValueError('Input either path or tensor') if img is not None: raise NotImplementedError _UpperCAmelCase = preprocess(Image.open(A) , target_image_size=2_56).to(self.device) _UpperCAmelCase = preprocess_vqgan(A) _UpperCAmelCase , *_UpperCAmelCase = self.vqgan.encode(A) return z def _lowerCamelCase ( self : List[str] , A : int) -> Dict: """simple docstring""" _UpperCAmelCase = self.latent.detach().requires_grad_() _UpperCAmelCase = base_latent + transform_vector if self.quantize: _UpperCAmelCase , *_UpperCAmelCase = self.vqgan.quantize(A) else: _UpperCAmelCase = trans_latent return self.vqgan.decode(A) def _lowerCamelCase ( self : Any , A : Dict , A : Dict , A : Optional[Any]=None) -> Any: """simple docstring""" _UpperCAmelCase = self.clip_preprocessor(text=A , images=A , return_tensors='pt' , padding=A) _UpperCAmelCase = self.clip(**A) _UpperCAmelCase = clip_outputs.logits_per_image if weights is not None: _UpperCAmelCase = similarity_logits * weights return similarity_logits.sum() def _lowerCamelCase ( self : Optional[int] , A : Dict , A : int , A : Tuple) -> str: """simple docstring""" _UpperCAmelCase = self._get_clip_similarity(pos_prompts['prompts'] , A , weights=(1 / pos_prompts['weights'])) if neg_prompts: _UpperCAmelCase = self._get_clip_similarity(neg_prompts['prompts'] , A , weights=neg_prompts['weights']) else: _UpperCAmelCase = torch.tensor([1] , device=self.device) _UpperCAmelCase = -torch.log(A) + torch.log(A) return loss def _lowerCamelCase ( self : Tuple , A : Optional[int] , A : List[Any] , A : Optional[int]) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = torch.randn_like(self.latent , requires_grad=A , device=self.device) _UpperCAmelCase = torch.optim.Adam([vector] , lr=self.lr) for i in range(self.iterations): optim.zero_grad() _UpperCAmelCase = self._add_vector(A) _UpperCAmelCase = loop_post_process(A) _UpperCAmelCase = self._get_CLIP_loss(A , A , A) print('CLIP loss' , A) if self.log: wandb.log({'CLIP Loss': clip_loss}) clip_loss.backward(retain_graph=A) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0]) else: yield vector def _lowerCamelCase ( self : Dict , A : Any , A : Optional[int] , A : str) -> Any: """simple docstring""" wandb.init(reinit=A , project='face-editor') wandb.config.update({'Positive Prompts': positive_prompts}) wandb.config.update({'Negative Prompts': negative_prompts}) wandb.config.update({'lr': self.lr, 'iterations': self.iterations}) if image_path: _UpperCAmelCase = Image.open(A) _UpperCAmelCase = image.resize((2_56, 2_56)) wandb.log('Original Image' , wandb.Image(A)) def _lowerCamelCase ( self : Dict , A : int) -> Dict: """simple docstring""" if not prompts: return [] _UpperCAmelCase = [] _UpperCAmelCase = [] if isinstance(A , A): _UpperCAmelCase = [prompt.strip() for prompt in prompts.split('|')] for prompt in prompts: if isinstance(A , (tuple, list)): _UpperCAmelCase = prompt[0] _UpperCAmelCase = float(prompt[1]) elif ":" in prompt: _UpperCAmelCase , _UpperCAmelCase = prompt.split(':') _UpperCAmelCase = float(A) else: _UpperCAmelCase = prompt _UpperCAmelCase = 1.0 processed_prompts.append(A) weights.append(A) return { "prompts": processed_prompts, "weights": torch.tensor(A , device=self.device), } def _lowerCamelCase ( self : Optional[int] , A : Union[str, Any] , A : Union[str, Any]=None , A : int=None , A : Optional[Any]=True , A : Dict=False , A : Union[str, Any]=True , A : Any=True , A : Any=None , ) -> Dict: """simple docstring""" if image_path: _UpperCAmelCase = self._get_latent(A) else: _UpperCAmelCase = torch.randn(self.latent_dim , device=self.device) if self.log: self._init_logging(A , A , A) assert pos_prompts, "You must provide at least one positive prompt." _UpperCAmelCase = self.process_prompts(A) _UpperCAmelCase = self.process_prompts(A) if save_final and save_path is None: _UpperCAmelCase = os.path.join('./outputs/' , '_'.join(pos_prompts['prompts'])) if not os.path.exists(A): os.makedirs(A) else: _UpperCAmelCase = save_path + '_' + get_timestamp() os.makedirs(A) _UpperCAmelCase = save_path _UpperCAmelCase = self.vqgan.decode(self.latent)[0] if show_intermediate: print('Original Image') show_pil(custom_to_pil(A)) _UpperCAmelCase = loop_post_process(A) for iter, transformed_img in enumerate(self._optimize_CLIP(A , A , A)): if show_intermediate: show_pil(A) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}.png")) if self.log: wandb.log({'Image': wandb.Image(A)}) if show_final: show_pil(A) if save_final: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}_final.png"))
290
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCAmelCase_ : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase_ : Optional[int] = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} UpperCAmelCase_ : str = { '''tokenizer_file''': { '''EleutherAI/gpt-neox-20b''': '''https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json''', }, } UpperCAmelCase_ : List[Any] = { '''gpt-neox-20b''': 2048, } class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): snake_case__ : List[str] = VOCAB_FILES_NAMES snake_case__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP snake_case__ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ : List[str] = ['input_ids', 'attention_mask'] def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , SCREAMING_SNAKE_CASE__ : Any=None , SCREAMING_SNAKE_CASE__ : List[str]=None , SCREAMING_SNAKE_CASE__ : List[str]="<|endoftext|>" , SCREAMING_SNAKE_CASE__ : Optional[int]="<|endoftext|>" , SCREAMING_SNAKE_CASE__ : Dict="<|endoftext|>" , SCREAMING_SNAKE_CASE__ : Tuple=False , **SCREAMING_SNAKE_CASE__ : Tuple , ) -> str: super().__init__( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , tokenizer_file=SCREAMING_SNAKE_CASE__ , unk_token=SCREAMING_SNAKE_CASE__ , bos_token=SCREAMING_SNAKE_CASE__ , eos_token=SCREAMING_SNAKE_CASE__ , add_prefix_space=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) a_ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , SCREAMING_SNAKE_CASE__ ) != add_prefix_space: a_ : Union[str, Any] = getattr(SCREAMING_SNAKE_CASE__ , pre_tok_state.pop('type' ) ) a_ : Dict = add_prefix_space a_ : Any = pre_tok_class(**SCREAMING_SNAKE_CASE__ ) a_ : Optional[int] = add_prefix_space def SCREAMING_SNAKE_CASE ( self : Any , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Tuple = None ) -> Optional[int]: a_ : List[str] = self._tokenizer.model.save(SCREAMING_SNAKE_CASE__ , name=SCREAMING_SNAKE_CASE__ ) return tuple(SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Dict , SCREAMING_SNAKE_CASE__ : str ) -> List[str]: a_ : Union[str, Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(SCREAMING_SNAKE_CASE__ , add_special_tokens=SCREAMING_SNAKE_CASE__ ) + [self.eos_token_id] ) if len(SCREAMING_SNAKE_CASE__ ) > self.model_max_length: a_ : Optional[int] = input_ids[-self.model_max_length :] return input_ids
32
'''simple docstring''' from math import pow def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): '''simple docstring''' if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count A_ : Optional[int] = int(pow(lowerCamelCase__ , lowerCamelCase__ ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n A_, A_ : int = backtrack( lowerCamelCase__ , lowerCamelCase__ , current_number + 1 , lowerCamelCase__ , lowerCamelCase__ ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. A_, A_ : int = backtrack( lowerCamelCase__ , lowerCamelCase__ , current_number + 1 , lowerCamelCase__ , lowerCamelCase__ ) return current_sum, solutions_count def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' if not (1 <= needed_sum <= 10_00 and 2 <= power <= 10): raise ValueError( """Invalid input\n""" """needed_sum must be between 1 and 1000, power between 2 and 10.""" ) return backtrack(lowerCamelCase__ , lowerCamelCase__ , 1 , 0 , 0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
206
0
'''simple docstring''' import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope="""session""" ) def _a ( ) -> str: """simple docstring""" __snake_case : str = 10 __snake_case : List[str] = datasets.Features( { """tokens""": datasets.Sequence(datasets.Value("""string""" ) ), """labels""": datasets.Sequence(datasets.ClassLabel(names=["""negative""", """positive"""] ) ), """answers""": datasets.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), """id""": datasets.Value("""int64""" ), } ) __snake_case : Tuple = datasets.Dataset.from_dict( { """tokens""": [["""foo"""] * 5] * n, """labels""": [[1] * 5] * n, """answers""": [{"""answer_start""": [97], """text""": ["""1976"""]}] * 10, """id""": list(range(A__ ) ), } , features=A__ , ) return dataset @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> List[Any]: """simple docstring""" __snake_case : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """file.arrow""" ) dataset.map(cache_file_name=A__ ) return filename # FILE_CONTENT + files __UpperCamelCase = "\\n Text data.\n Second line of data." @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Optional[Any]: """simple docstring""" __snake_case : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """file.txt""" __snake_case : Union[str, Any] = FILE_CONTENT with open(A__ , """w""" ) as f: f.write(A__ ) return filename @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Dict: """simple docstring""" import bza __snake_case : str = tmp_path_factory.mktemp("""data""" ) / """file.txt.bz2""" __snake_case : Tuple = bytes(A__ , """utf-8""" ) with bza.open(A__ , """wb""" ) as f: f.write(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> int: """simple docstring""" import gzip __snake_case : Dict = str(tmp_path_factory.mktemp("""data""" ) / """file.txt.gz""" ) __snake_case : int = bytes(A__ , """utf-8""" ) with gzip.open(A__ , """wb""" ) as f: f.write(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> List[Any]: """simple docstring""" if datasets.config.LZ4_AVAILABLE: import lza.frame __snake_case : Union[str, Any] = tmp_path_factory.mktemp("""data""" ) / """file.txt.lz4""" __snake_case : Optional[int] = bytes(A__ , """utf-8""" ) with lza.frame.open(A__ , """wb""" ) as f: f.write(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> List[str]: """simple docstring""" if datasets.config.PY7ZR_AVAILABLE: import pyazr __snake_case : Any = tmp_path_factory.mktemp("""data""" ) / """file.txt.7z""" with pyazr.SevenZipFile(A__ , """w""" ) as archive: archive.write(A__ , arcname=os.path.basename(A__ ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> str: """simple docstring""" import tarfile __snake_case : Any = tmp_path_factory.mktemp("""data""" ) / """file.txt.tar""" with tarfile.TarFile(A__ , """w""" ) as f: f.add(A__ , arcname=os.path.basename(A__ ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Tuple: """simple docstring""" import lzma __snake_case : Dict = tmp_path_factory.mktemp("""data""" ) / """file.txt.xz""" __snake_case : Dict = bytes(A__ , """utf-8""" ) with lzma.open(A__ , """wb""" ) as f: f.write(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> Any: """simple docstring""" import zipfile __snake_case : Tuple = tmp_path_factory.mktemp("""data""" ) / """file.txt.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.basename(A__ ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Tuple: """simple docstring""" if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd __snake_case : str = tmp_path_factory.mktemp("""data""" ) / """file.txt.zst""" __snake_case : Tuple = bytes(A__ , """utf-8""" ) with zstd.open(A__ , """wb""" ) as f: f.write(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Union[str, Any]: """simple docstring""" __snake_case : Any = tmp_path_factory.mktemp("""data""" ) / """file.xml""" __snake_case : Tuple = textwrap.dedent( """\ <?xml version=\"1.0\" encoding=\"UTF-8\" ?> <tmx version=\"1.4\"> <header segtype=\"sentence\" srclang=\"ca\" /> <body> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv> </tu> </body> </tmx>""" ) with open(A__ , """w""" ) as f: f.write(A__ ) return filename __UpperCamelCase = [ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] __UpperCamelCase = [ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] __UpperCamelCase = { "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } __UpperCamelCase = [ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] __UpperCamelCase = [ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope="""session""" ) def _a ( ) -> Optional[Any]: """simple docstring""" return DATA_DICT_OF_LISTS @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> int: """simple docstring""" __snake_case : Any = datasets.Dataset.from_dict(A__ ) __snake_case : int = str(tmp_path_factory.mktemp("""data""" ) / """dataset.arrow""" ) dataset.map(cache_file_name=A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> int: """simple docstring""" __snake_case : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset.sqlite""" ) with contextlib.closing(sqlitea.connect(A__ ) ) as con: __snake_case : Any = con.cursor() cur.execute("""CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)""" ) for item in DATA: cur.execute("""INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)""" , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Tuple: """simple docstring""" __snake_case : Any = str(tmp_path_factory.mktemp("""data""" ) / """dataset.csv""" ) with open(A__ , """w""" , newline="""""" ) as f: __snake_case : Union[str, Any] = csv.DictWriter(A__ , fieldnames=["""col_1""", """col_2""", """col_3"""] ) writer.writeheader() for item in DATA: writer.writerow(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> List[str]: """simple docstring""" __snake_case : str = str(tmp_path_factory.mktemp("""data""" ) / """dataset2.csv""" ) with open(A__ , """w""" , newline="""""" ) as f: __snake_case : Optional[int] = csv.DictWriter(A__ , fieldnames=["""col_1""", """col_2""", """col_3"""] ) writer.writeheader() for item in DATA: writer.writerow(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> List[Any]: """simple docstring""" import bza __snake_case : Dict = tmp_path_factory.mktemp("""data""" ) / """dataset.csv.bz2""" with open(A__ , """rb""" ) as f: __snake_case : Optional[int] = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(A__ , """wb""" ) as f: f.write(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: """simple docstring""" __snake_case : List[Any] = tmp_path_factory.mktemp("""data""" ) / """dataset.csv.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.basename(A__ ) ) f.write(A__ , arcname=os.path.basename(A__ ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: """simple docstring""" __snake_case : int = tmp_path_factory.mktemp("""data""" ) / """dataset.csv.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.basename(csv_path.replace(""".csv""" , """.CSV""" ) ) ) f.write(A__ , arcname=os.path.basename(csva_path.replace(""".csv""" , """.CSV""" ) ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Any: """simple docstring""" __snake_case : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.csv.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.join("""main_dir""" , os.path.basename(A__ ) ) ) f.write(A__ , arcname=os.path.join("""main_dir""" , os.path.basename(A__ ) ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Optional[int]: """simple docstring""" __snake_case : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset.parquet""" ) __snake_case : str = pa.schema( { """col_1""": pa.string(), """col_2""": pa.intaa(), """col_3""": pa.floataa(), } ) with open(A__ , """wb""" ) as f: __snake_case : Optional[int] = pq.ParquetWriter(A__ , schema=A__ ) __snake_case : str = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(A__ ) )] for k in DATA[0]} , schema=A__ ) writer.write_table(A__ ) writer.close() return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> str: """simple docstring""" __snake_case : str = str(tmp_path_factory.mktemp("""data""" ) / """dataset.json""" ) __snake_case : List[str] = {"""data""": DATA} with open(A__ , """w""" ) as f: json.dump(A__ , A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> int: """simple docstring""" __snake_case : Optional[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.json""" ) __snake_case : Tuple = {"""data""": DATA_DICT_OF_LISTS} with open(A__ , """w""" ) as f: json.dump(A__ , A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Union[str, Any]: """simple docstring""" __snake_case : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl""" ) with open(A__ , """w""" ) as f: for item in DATA: f.write(json.dumps(A__ ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Optional[Any]: """simple docstring""" __snake_case : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset2.jsonl""" ) with open(A__ , """w""" ) as f: for item in DATA: f.write(json.dumps(A__ ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Optional[Any]: """simple docstring""" __snake_case : Any = str(tmp_path_factory.mktemp("""data""" ) / """dataset_312.jsonl""" ) with open(A__ , """w""" ) as f: for item in DATA_312: f.write(json.dumps(A__ ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> List[Any]: """simple docstring""" __snake_case : Any = str(tmp_path_factory.mktemp("""data""" ) / """dataset-str.jsonl""" ) with open(A__ , """w""" ) as f: for item in DATA_STR: f.write(json.dumps(A__ ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: """simple docstring""" import gzip __snake_case : List[str] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.txt.gz""" ) with open(A__ , """rb""" ) as orig_file: with gzip.open(A__ , """wb""" ) as zipped_file: zipped_file.writelines(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> Tuple: """simple docstring""" import gzip __snake_case : Optional[int] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.gz""" ) with open(A__ , """rb""" ) as orig_file: with gzip.open(A__ , """wb""" ) as zipped_file: zipped_file.writelines(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> int: """simple docstring""" __snake_case : List[str] = tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.basename(A__ ) ) f.write(A__ , arcname=os.path.basename(A__ ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> str: """simple docstring""" __snake_case : Tuple = tmp_path_factory.mktemp("""data""" ) / """dataset_nested.jsonl.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.join("""nested""" , os.path.basename(A__ ) ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[Any]: """simple docstring""" __snake_case : Optional[Any] = tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.jsonl.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.join("""main_dir""" , os.path.basename(A__ ) ) ) f.write(A__ , arcname=os.path.join("""main_dir""" , os.path.basename(A__ ) ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Dict: """simple docstring""" __snake_case : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.tar""" with tarfile.TarFile(A__ , """w""" ) as f: f.add(A__ , arcname=os.path.basename(A__ ) ) f.add(A__ , arcname=os.path.basename(A__ ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Dict: """simple docstring""" __snake_case : Union[str, Any] = tmp_path_factory.mktemp("""data""" ) / """dataset_nested.jsonl.tar""" with tarfile.TarFile(A__ , """w""" ) as f: f.add(A__ , arcname=os.path.join("""nested""" , os.path.basename(A__ ) ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> int: """simple docstring""" __snake_case : Optional[Any] = ["""0""", """1""", """2""", """3"""] __snake_case : Union[str, Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.txt""" ) with open(A__ , """w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Any: """simple docstring""" __snake_case : Union[str, Any] = ["""0""", """1""", """2""", """3"""] __snake_case : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset2.txt""" ) with open(A__ , """w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Union[str, Any]: """simple docstring""" __snake_case : str = ["""0""", """1""", """2""", """3"""] __snake_case : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """dataset.abc""" with open(A__ , """w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[Any]: """simple docstring""" __snake_case : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """dataset.text.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.basename(A__ ) ) f.write(A__ , arcname=os.path.basename(A__ ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Dict: """simple docstring""" __snake_case : List[Any] = tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.text.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.join("""main_dir""" , os.path.basename(A__ ) ) ) f.write(A__ , arcname=os.path.join("""main_dir""" , os.path.basename(A__ ) ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: """simple docstring""" __snake_case : List[str] = tmp_path_factory.mktemp("""data""" ) / """dataset.ext.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.basename("""unsupported.ext""" ) ) f.write(A__ , arcname=os.path.basename("""unsupported_2.ext""" ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> Any: """simple docstring""" __snake_case : Optional[Any] = """\n""".join(["""First""", """Second\u2029with Unicode new line""", """Third"""] ) __snake_case : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset_with_unicode_new_lines.txt""" ) with open(A__ , """w""" , encoding="""utf-8""" ) as f: f.write(A__ ) return path @pytest.fixture(scope="""session""" ) def _a ( ) -> int: """simple docstring""" return os.path.join("""tests""" , """features""" , """data""" , """test_image_rgb.jpg""" ) @pytest.fixture(scope="""session""" ) def _a ( ) -> Dict: """simple docstring""" return os.path.join("""tests""" , """features""" , """data""" , """test_audio_44100.wav""" ) @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase , _lowerCamelCase ) -> str: """simple docstring""" __snake_case : List[Any] = tmp_path_factory.mktemp("""data""" ) / """dataset.img.zip""" with zipfile.ZipFile(A__ , """w""" ) as f: f.write(A__ , arcname=os.path.basename(A__ ) ) f.write(A__ , arcname=os.path.basename(A__ ).replace(""".jpg""" , """2.jpg""" ) ) return path @pytest.fixture(scope="""session""" ) def _a ( _lowerCamelCase ) -> int: """simple docstring""" __snake_case : List[Any] = tmp_path_factory.mktemp("""data_dir""" ) (data_dir / "subdir").mkdir() with open(data_dir / """subdir""" / """train.txt""" , """w""" ) as f: f.write("""foo\n""" * 10 ) with open(data_dir / """subdir""" / """test.txt""" , """w""" ) as f: f.write("""bar\n""" * 10 ) # hidden file with open(data_dir / """subdir""" / """.test.txt""" , """w""" ) as f: f.write("""bar\n""" * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / """.subdir""" / """train.txt""" , """w""" ) as f: f.write("""foo\n""" * 10 ) with open(data_dir / """.subdir""" / """test.txt""" , """w""" ) as f: f.write("""bar\n""" * 10 ) return data_dir
350
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class _A ( unittest.TestCase ): @slow def lowercase__ ( self : List[str] ) -> int: """simple docstring""" __snake_case : List[Any] = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) __snake_case : Tuple = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" __snake_case : List[str] = model(__magic_name__ )["""last_hidden_state"""] __snake_case : Any = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , __magic_name__ ) # compare the actual values for a slice. __snake_case : str = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
13
0
'''simple docstring''' from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def __lowerCamelCase ( __snake_case : Sequence[float], __snake_case : int, __snake_case : int ) -> tuple[int | None, int | None, float]: """simple docstring""" if not arr: return None, None, 0 if low == high: return low, high, arr[low] A__ : List[Any] =(low + high) // 2 A__ , A__ , A__ : Optional[Any] =max_subarray(__snake_case, __snake_case, __snake_case ) A__ , A__ , A__ : Union[str, Any] =max_subarray(__snake_case, mid + 1, __snake_case ) A__ , A__ , A__ : Optional[Any] =max_cross_sum(__snake_case, __snake_case, __snake_case, __snake_case ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def __lowerCamelCase ( __snake_case : Sequence[float], __snake_case : int, __snake_case : int, __snake_case : int ) -> tuple[int, int, float]: """simple docstring""" A__ , A__ : str =float("""-inf""" ), -1 A__ , A__ : Optional[Any] =float("""-inf""" ), -1 A__ : int | float =0 for i in range(__snake_case, low - 1, -1 ): summ += arr[i] if summ > left_sum: A__ : List[Any] =summ A__ : List[Any] =i A__ : Dict =0 for i in range(mid + 1, high + 1 ): summ += arr[i] if summ > right_sum: A__ : Dict =summ A__ : List[str] =i return max_left, max_right, (left_sum + right_sum) def __lowerCamelCase ( __snake_case : int ) -> float: """simple docstring""" A__ : str =[randint(1, __snake_case ) for _ in range(__snake_case )] A__ : Optional[int] =time.time() max_subarray(__snake_case, 0, input_size - 1 ) A__ : str =time.time() return end - start def __lowerCamelCase ( ) -> None: """simple docstring""" A__ : Dict =[10, 100, 1_000, 10_000, 50_000, 100_000, 200_000, 300_000, 400_000, 500_000] A__ : Any =[time_max_subarray(__snake_case ) for input_size in input_sizes] print("""No of Inputs\t\tTime Taken""" ) for input_size, runtime in zip(__snake_case, __snake_case ): print(__snake_case, """\t\t""", __snake_case ) plt.plot(__snake_case, __snake_case ) plt.xlabel("""Number of Inputs""" ) plt.ylabel("""Time taken in seconds""" ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
134
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case : Optional[Any] = {'configuration_xglm': ['XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XGLMConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : int = ['XGLMTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : List[Any] = ['XGLMTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[Any] = [ 'XGLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XGLMForCausalLM', 'XGLMModel', 'XGLMPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : str = [ 'FlaxXGLMForCausalLM', 'FlaxXGLMModel', 'FlaxXGLMPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Optional[Any] = [ 'TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXGLMForCausalLM', 'TFXGLMModel', 'TFXGLMPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure)
134
1
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() lowerCAmelCase__ : Any =logging.get_logger(__name__) lowerCAmelCase__ : Dict ={ '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn.grep_linear''': '''encoder.layers.*.attention.gru_rel_pos_linear''', '''self_attn.relative_attention_bias''': '''encoder.layers.*.attention.rel_attn_embed''', '''self_attn.grep_a''': '''encoder.layers.*.attention.gru_rel_pos_const''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''ctc_proj''', '''mask_emb''': '''masked_spec_embed''', } lowerCAmelCase__ : List[str] =[ '''ctc_proj''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def __lowercase ( a__ , a__ , a__ , a__ , a__ ) -> List[Any]: for attribute in key.split('.' ): __SCREAMING_SNAKE_CASE = getattr(a__ , a__ ) if weight_type is not None: __SCREAMING_SNAKE_CASE = getattr(a__ , a__ ).shape else: __SCREAMING_SNAKE_CASE = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __SCREAMING_SNAKE_CASE = value elif weight_type == "weight_g": __SCREAMING_SNAKE_CASE = value elif weight_type == "weight_v": __SCREAMING_SNAKE_CASE = value elif weight_type == "bias": __SCREAMING_SNAKE_CASE = value else: __SCREAMING_SNAKE_CASE = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def __lowercase ( a__ , a__ ) -> int: __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = fairseq_model.state_dict() __SCREAMING_SNAKE_CASE = hf_model.feature_extractor for name, value in fairseq_dict.items(): __SCREAMING_SNAKE_CASE = False if "conv_layers" in name: load_conv_layer( a__ , a__ , a__ , a__ , hf_model.config.feat_extract_norm == 'group' , ) __SCREAMING_SNAKE_CASE = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: __SCREAMING_SNAKE_CASE = True if "*" in mapped_key: __SCREAMING_SNAKE_CASE = name.split(a__ )[0].split('.' )[-2] __SCREAMING_SNAKE_CASE = mapped_key.replace('*' , a__ ) if "weight_g" in name: __SCREAMING_SNAKE_CASE = 'weight_g' elif "weight_v" in name: __SCREAMING_SNAKE_CASE = 'weight_v' elif "bias" in name and "relative_attention_bias" not in name: __SCREAMING_SNAKE_CASE = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj __SCREAMING_SNAKE_CASE = 'weight' else: __SCREAMING_SNAKE_CASE = None set_recursively(a__ , a__ , a__ , a__ , a__ ) continue if not is_used: unused_weights.append(a__ ) logger.warning(f"""Unused weights: {unused_weights}""" ) def __lowercase ( a__ , a__ , a__ , a__ , a__ ) -> Dict: __SCREAMING_SNAKE_CASE = full_name.split('conv_layers.' )[-1] __SCREAMING_SNAKE_CASE = name.split('.' ) __SCREAMING_SNAKE_CASE = int(items[0] ) __SCREAMING_SNAKE_CASE = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __SCREAMING_SNAKE_CASE = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __SCREAMING_SNAKE_CASE = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __SCREAMING_SNAKE_CASE = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __SCREAMING_SNAKE_CASE = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(a__ ) @torch.no_grad() def __lowercase ( a__ , a__ , a__=None ) -> Union[str, Any]: # load the pre-trained checkpoints __SCREAMING_SNAKE_CASE = torch.load(a__ ) __SCREAMING_SNAKE_CASE = WavLMConfigOrig(checkpoint['cfg'] ) __SCREAMING_SNAKE_CASE = WavLMOrig(a__ ) model.load_state_dict(checkpoint['model'] ) model.eval() if config_path is not None: __SCREAMING_SNAKE_CASE = WavLMConfig.from_pretrained(a__ ) else: __SCREAMING_SNAKE_CASE = WavLMConfig() __SCREAMING_SNAKE_CASE = WavLMModel(a__ ) recursively_load_weights(a__ , a__ ) hf_wavlm.save_pretrained(a__ ) if __name__ == "__main__": lowerCAmelCase__ : List[Any] =argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') lowerCAmelCase__ : Optional[int] =parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
118
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class UpperCAmelCase_ ( UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): '''simple docstring''' UpperCamelCase__ : List[str] = CycleDiffusionPipeline UpperCamelCase__ : int = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { '''negative_prompt''', '''height''', '''width''', '''negative_prompt_embeds''', } UpperCamelCase__ : Union[str, Any] = PipelineTesterMixin.required_optional_params - {'''latents'''} UpperCamelCase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''source_prompt'''} ) UpperCamelCase__ : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS UpperCamelCase__ : str = IMAGE_TO_IMAGE_IMAGE_PARAMS def _A ( self ): '''simple docstring''' torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) __SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='scaled_linear' , num_train_timesteps=1_000 , clip_sample=_A , set_alpha_to_one=_A , ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) __SCREAMING_SNAKE_CASE = CLIPTextModel(_A ) __SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __SCREAMING_SNAKE_CASE = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _A ( self , _A , _A=0 ): '''simple docstring''' __SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A ) ).to(_A ) __SCREAMING_SNAKE_CASE = image / 2 + 0.5 if str(_A ).startswith('mps' ): __SCREAMING_SNAKE_CASE = torch.manual_seed(_A ) else: __SCREAMING_SNAKE_CASE = torch.Generator(device=_A ).manual_seed(_A ) __SCREAMING_SNAKE_CASE = { 'prompt': 'An astronaut riding an elephant', 'source_prompt': 'An astronaut riding a horse', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'eta': 0.1, 'strength': 0.8, 'guidance_scale': 3, 'source_guidance_scale': 1, 'output_type': 'numpy', } return inputs def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 'cpu' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE = self.get_dummy_components() __SCREAMING_SNAKE_CASE = CycleDiffusionPipeline(**_A ) __SCREAMING_SNAKE_CASE = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __SCREAMING_SNAKE_CASE = self.get_dummy_inputs(_A ) __SCREAMING_SNAKE_CASE = pipe(**_A ) __SCREAMING_SNAKE_CASE = output.images __SCREAMING_SNAKE_CASE = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE = np.array([0.4_4_5_9, 0.4_9_4_3, 0.4_5_4_4, 0.6_6_4_3, 0.5_4_7_4, 0.4_3_2_7, 0.5_7_0_1, 0.5_9_5_9, 0.5_1_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.get_dummy_components() for name, module in components.items(): if hasattr(_A , 'half' ): __SCREAMING_SNAKE_CASE = module.half() __SCREAMING_SNAKE_CASE = CycleDiffusionPipeline(**_A ) __SCREAMING_SNAKE_CASE = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __SCREAMING_SNAKE_CASE = self.get_dummy_inputs(_A ) __SCREAMING_SNAKE_CASE = pipe(**_A ) __SCREAMING_SNAKE_CASE = output.images __SCREAMING_SNAKE_CASE = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE = np.array([0.3_5_0_6, 0.4_5_4_3, 0.4_4_6, 0.4_5_7_5, 0.5_1_9_5, 0.4_1_5_5, 0.5_2_7_3, 0.5_1_8, 0.4_1_1_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def _A ( self ): '''simple docstring''' return super().test_save_load_local() @unittest.skip('non-deterministic pipeline' ) def _A ( self ): '''simple docstring''' return super().test_inference_batch_single_identical() @skip_mps def _A ( self ): '''simple docstring''' return super().test_dict_tuple_outputs_equivalent() @skip_mps def _A ( self ): '''simple docstring''' return super().test_save_load_optional_components() @skip_mps def _A ( self ): '''simple docstring''' return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _A ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) __SCREAMING_SNAKE_CASE = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy' ) __SCREAMING_SNAKE_CASE = init_image.resize((512, 512) ) __SCREAMING_SNAKE_CASE = 'CompVis/stable-diffusion-v1-4' __SCREAMING_SNAKE_CASE = DDIMScheduler.from_pretrained(_A , subfolder='scheduler' ) __SCREAMING_SNAKE_CASE = CycleDiffusionPipeline.from_pretrained( _A , scheduler=_A , safety_checker=_A , torch_dtype=torch.floataa , revision='fp16' ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE = 'A black colored car' __SCREAMING_SNAKE_CASE = 'A blue colored car' __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=_A , source_prompt=_A , image=_A , num_inference_steps=100 , eta=0.1 , strength=0.8_5 , guidance_scale=3 , source_guidance_scale=1 , generator=_A , output_type='np' , ) __SCREAMING_SNAKE_CASE = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5e-1 def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) __SCREAMING_SNAKE_CASE = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy' ) __SCREAMING_SNAKE_CASE = init_image.resize((512, 512) ) __SCREAMING_SNAKE_CASE = 'CompVis/stable-diffusion-v1-4' __SCREAMING_SNAKE_CASE = DDIMScheduler.from_pretrained(_A , subfolder='scheduler' ) __SCREAMING_SNAKE_CASE = CycleDiffusionPipeline.from_pretrained(_A , scheduler=_A , safety_checker=_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE = 'A black colored car' __SCREAMING_SNAKE_CASE = 'A blue colored car' __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=_A , source_prompt=_A , image=_A , num_inference_steps=100 , eta=0.1 , strength=0.8_5 , guidance_scale=3 , source_guidance_scale=1 , generator=_A , output_type='np' , ) __SCREAMING_SNAKE_CASE = output.images assert np.abs(image - expected_image ).max() < 2e-2
118
1
"""simple docstring""" from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration UpperCAmelCase : Optional[int] = HfArgumentParser(InitializationArguments) UpperCAmelCase : List[str] = parser.parse_args() # Load codeparrot tokenizer trained for Python code tokenization UpperCAmelCase : Optional[int] = AutoTokenizer.from_pretrained(args.tokenizer_name) # Config: "scale_attn_by_layer_idx" and "reorder_and_upcast_attn" are Mistral stability tweaks UpperCAmelCase : Optional[int] = { 'vocab_size': len(tokenizer), 'scale_attn_by_inverse_layer_idx': True, 'reorder_and_upcast_attn': True, } # Load model config (GPT-2 large in this case) UpperCAmelCase : List[Any] = AutoConfig.from_pretrained(args.config_name, **config_kwargs) # Initialize new model with config UpperCAmelCase : Optional[int] = AutoModelForCausalLM.from_config(config) # Save model to the hub model.save_pretrained(args.model_name, push_to_hub=args.push_to_hub)
115
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase : Any = { 'configuration_blenderbot': [ 'BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlenderbotConfig', 'BlenderbotOnnxConfig', ], 'tokenization_blenderbot': ['BlenderbotTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[int] = ['BlenderbotTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[Any] = [ 'BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BlenderbotForCausalLM', 'BlenderbotForConditionalGeneration', 'BlenderbotModel', 'BlenderbotPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[int] = [ 'TFBlenderbotForConditionalGeneration', 'TFBlenderbotModel', 'TFBlenderbotPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Optional[int] = [ 'FlaxBlenderbotForConditionalGeneration', 'FlaxBlenderbotModel', 'FlaxBlenderbotPreTrainedModel', ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys UpperCAmelCase : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
115
1
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class snake_case_ ( __lowercase ,__lowercase ,unittest.TestCase ): A_ = IFInpaintingSuperResolutionPipeline A_ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'width', 'height'} A_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'original_image'} ) A_ = PipelineTesterMixin.required_optional_params - {'latents'} def UpperCAmelCase__ ( self : str )->List[Any]: '''simple docstring''' return self._get_superresolution_dummy_components() def UpperCAmelCase__ ( self : int , _snake_case : Tuple , _snake_case : List[str]=0 )->Tuple: '''simple docstring''' if str(_snake_case ).startswith("""mps""" ): __lowerCAmelCase : List[str] = torch.manual_seed(_snake_case ) else: __lowerCAmelCase : int = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) __lowerCAmelCase : Tuple = floats_tensor((1, 3, 16, 16) , rng=random.Random(_snake_case ) ).to(_snake_case ) __lowerCAmelCase : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(_snake_case ) ).to(_snake_case ) __lowerCAmelCase : Dict = floats_tensor((1, 3, 32, 32) , rng=random.Random(_snake_case ) ).to(_snake_case ) __lowerCAmelCase : List[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """original_image""": original_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def UpperCAmelCase__ ( self : int )->Dict: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCAmelCase__ ( self : Any )->Dict: '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def UpperCAmelCase__ ( self : Optional[int] )->str: '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCAmelCase__ ( self : Any )->Optional[Any]: '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCAmelCase__ ( self : Dict )->List[str]: '''simple docstring''' self._test_save_load_local() def UpperCAmelCase__ ( self : str )->List[Any]: '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
232
_UpperCAmelCase = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} _UpperCAmelCase = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :dict[int, list[int]] , SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :list[bool] ) -> list[int]: __lowerCAmelCase : str = True __lowerCAmelCase : str = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) order.append(SCREAMING_SNAKE_CASE ) return order def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :dict[int, list[int]] , SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :list[bool] ) -> list[int]: __lowerCAmelCase : Optional[Any] = True __lowerCAmelCase : Union[str, Any] = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return component def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :dict[int, list[int]] ) -> list[list[int]]: __lowerCAmelCase : Optional[Any] = len(SCREAMING_SNAKE_CASE ) * [False] __lowerCAmelCase : dict[int, list[int]] = {vert: [] for vert in range(len(SCREAMING_SNAKE_CASE ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Tuple = [] for i, was_visited in enumerate(SCREAMING_SNAKE_CASE ): if not was_visited: order += topology_sort(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = [] __lowerCAmelCase : Any = len(SCREAMING_SNAKE_CASE ) * [False] for i in range(len(SCREAMING_SNAKE_CASE ) ): __lowerCAmelCase : Optional[int] = order[len(SCREAMING_SNAKE_CASE ) - i - 1] if not visited[vert]: __lowerCAmelCase : Any = find_components(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) components_list.append(SCREAMING_SNAKE_CASE ) return components_list
232
1
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = CodeGenTokenizer a__ = CodeGenTokenizerFast a__ = True a__ = {"""add_prefix_space""": True} a__ = False def lowerCamelCase_ ( self) -> Any: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt a__: int = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] a__: List[str] = dict(zip(lowercase , range(len(lowercase)))) a__: Optional[int] = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] a__: Tuple = {'unk_token': '<unk>'} a__: Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) a__: Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file']) with open(self.vocab_file , 'w' , encoding='utf-8') as fp: fp.write(json.dumps(lowercase) + '\n') with open(self.merges_file , 'w' , encoding='utf-8') as fp: fp.write('\n'.join(lowercase)) def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' kwargs.update(self.special_tokens_map) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **lowercase) def lowerCamelCase_ ( self , **lowercase) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **lowercase) def lowerCamelCase_ ( self , lowercase) -> Optional[int]: '''simple docstring''' a__: List[str] = 'lower newer' a__: str = 'lower newer' return input_text, output_text def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: List[Any] = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) a__: Union[str, Any] = 'lower newer' a__: Any = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] a__: Tuple = tokenizer.tokenize(lowercase , add_prefix_space=lowercase) self.assertListEqual(lowercase , lowercase) a__: int = tokens + [tokenizer.unk_token] a__: Any = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase) , lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' if not self.test_rust_tokenizer: return a__: Union[str, Any] = self.get_tokenizer() a__: int = self.get_rust_tokenizer(add_prefix_space=lowercase) a__: Any = 'lower newer' # Testing tokenization a__: Any = tokenizer.tokenize(lowercase , add_prefix_space=lowercase) a__: Tuple = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) # Testing conversion to ids without special tokens a__: Optional[Any] = tokenizer.encode(lowercase , add_special_tokens=lowercase , add_prefix_space=lowercase) a__: Dict = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) # Testing conversion to ids with special tokens a__: Dict = self.get_rust_tokenizer(add_prefix_space=lowercase) a__: Tuple = tokenizer.encode(lowercase , add_prefix_space=lowercase) a__: Tuple = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) # Testing the unknown token a__: Any = tokens + [rust_tokenizer.unk_token] a__: Any = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(lowercase) , lowercase) def lowerCamelCase_ ( self , *lowercase , **lowercase) -> List[Any]: '''simple docstring''' pass def lowerCamelCase_ ( self , lowercase=15) -> int: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})'): a__: Optional[Any] = self.rust_tokenizer_class.from_pretrained(lowercase , **lowercase) # Simple input a__: Any = 'This is a simple input' a__: str = ['This is a simple input 1', 'This is a simple input 2'] a__: Any = ('This is a simple input', 'This is a pair') a__: int = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(lowercase , tokenizer_r.encode , lowercase , max_length=lowercase , padding='max_length') # Simple input self.assertRaises(lowercase , tokenizer_r.encode_plus , lowercase , max_length=lowercase , padding='max_length') # Simple input self.assertRaises( lowercase , tokenizer_r.batch_encode_plus , lowercase , max_length=lowercase , padding='max_length' , ) # Pair input self.assertRaises(lowercase , tokenizer_r.encode , lowercase , max_length=lowercase , padding='max_length') # Pair input self.assertRaises(lowercase , tokenizer_r.encode_plus , lowercase , max_length=lowercase , padding='max_length') # Pair input self.assertRaises( lowercase , tokenizer_r.batch_encode_plus , lowercase , max_length=lowercase , padding='max_length' , ) def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: str = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>') # Simple input a__: str = 'This is a simple input' a__: List[Any] = ['This is a simple input looooooooong', 'This is a simple input'] a__: int = ('This is a simple input', 'This is a pair') a__: str = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] a__: List[Any] = tokenizer.pad_token_id a__: Optional[Any] = tokenizer(lowercase , padding='max_length' , max_length=30 , return_tensors='np') a__: Union[str, Any] = tokenizer(lowercase , padding=lowercase , truncate=lowercase , return_tensors='np') a__: Union[str, Any] = tokenizer(*lowercase , padding='max_length' , max_length=60 , return_tensors='np') a__: Optional[Any] = tokenizer(lowercase , padding=lowercase , truncate=lowercase , return_tensors='np') # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 30) self.assertTrue(pad_token_id in out_s['input_ids']) self.assertTrue(0 in out_s['attention_mask']) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 33) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0]) self.assertFalse(0 in out_sa['attention_mask'][0]) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1]) self.assertTrue(0 in out_sa['attention_mask'][1]) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 60) self.assertTrue(pad_token_id in out_p['input_ids']) self.assertTrue(0 in out_p['attention_mask']) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 52) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0]) self.assertFalse(0 in out_pa['attention_mask'][0]) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1]) self.assertTrue(0 in out_pa['attention_mask'][1]) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Any = '$$$' a__: Union[str, Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=lowercase , add_bos_token=lowercase) a__: List[str] = 'This is a simple input' a__: List[Any] = ['This is a simple input 1', 'This is a simple input 2'] a__: Tuple = tokenizer.bos_token_id a__: Optional[Any] = tokenizer(lowercase) a__: Dict = tokenizer(lowercase) self.assertEqual(out_s.input_ids[0] , lowercase) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids)) a__: str = tokenizer.decode(out_s.input_ids) a__: List[str] = tokenizer.batch_decode(out_sa.input_ids) self.assertEqual(decode_s.split()[0] , lowercase) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa)) @slow def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = CodeGenTokenizer.from_pretrained('Salesforce/codegen-350M-mono') a__: int = '\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#' a__: Tuple = '\nif len_a > len_b: result = a\nelse: result = b' a__: Dict = tokenizer.encode(lowercase) a__: Union[str, Any] = ['^#', re.escape('<|endoftext|>'), '^\'\'\'', '^"""', '\n\n\n'] a__: List[str] = tokenizer.decode(lowercase , truncate_before_pattern=lowercase) self.assertEqual(lowercase , lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' pass
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'edbeeching/decision-transformer-gym-hopper-medium': ( 'https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class __snake_case ( __lowerCAmelCase ): a__ = """decision_transformer""" a__ = ["""past_key_values"""] a__ = { """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=17 , lowercase=4 , lowercase=1_28 , lowercase=40_96 , lowercase=True , lowercase=1 , lowercase=10_24 , lowercase=3 , lowercase=1 , lowercase=None , lowercase="relu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.1 , lowercase=1e-5 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=5_02_56 , lowercase=5_02_56 , lowercase=False , lowercase=False , **lowercase , ) -> Tuple: '''simple docstring''' a__: List[str] = state_dim a__: int = act_dim a__: List[Any] = hidden_size a__: List[str] = max_ep_len a__: List[Any] = action_tanh a__: Optional[Any] = vocab_size a__: Tuple = n_positions a__: Dict = n_layer a__: Optional[int] = n_head a__: Optional[int] = n_inner a__: Any = activation_function a__: Union[str, Any] = resid_pdrop a__: Any = embd_pdrop a__: Any = attn_pdrop a__: List[Any] = layer_norm_epsilon a__: Optional[Any] = initializer_range a__: Any = scale_attn_weights a__: Dict = use_cache a__: Optional[int] = scale_attn_by_inverse_layer_idx a__: List[str] = reorder_and_upcast_attn a__: Any = bos_token_id a__: int = eos_token_id super().__init__(bos_token_id=lowercase , eos_token_id=lowercase , **lowercase)
290
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging _UpperCamelCase: List[Any] = logging.get_logger(__name__) _UpperCamelCase: Optional[int] = '▁' _UpperCamelCase: List[Any] = {'vocab_file': 'sentencepiece.bpe.model'} _UpperCamelCase: Any = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), } } _UpperCamelCase: List[str] = { 'facebook/mbart-large-en-ro': 1_0_2_4, 'facebook/mbart-large-cc25': 1_0_2_4, } # fmt: off _UpperCamelCase: Tuple = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class a__ ( SCREAMING_SNAKE_CASE__ ): _lowerCamelCase = VOCAB_FILES_NAMES _lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase = ['input_ids', 'attention_mask'] _lowerCamelCase = [] _lowerCamelCase = [] def __init__( self : List[str], lowerCAmelCase : List[Any], lowerCAmelCase : Union[str, Any]="<s>", lowerCAmelCase : Union[str, Any]="</s>", lowerCAmelCase : List[Any]="</s>", lowerCAmelCase : str="<s>", lowerCAmelCase : Tuple="<unk>", lowerCAmelCase : str="<pad>", lowerCAmelCase : List[str]="<mask>", lowerCAmelCase : Any=None, lowerCAmelCase : int=None, lowerCAmelCase : Tuple=None, lowerCAmelCase : Optional[Dict[str, Any]] = None, lowerCAmelCase : List[str]=None, **lowerCAmelCase : Tuple, ) -> int: # Mask token behave like a normal word, i.e. include the space before it lowercase : Union[str, Any] = AddedToken(lowerCAmelCase, lstrip=lowerCAmelCase, rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase, lowerCAmelCase ) else mask_token lowercase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCAmelCase, eos_token=lowerCAmelCase, unk_token=lowerCAmelCase, sep_token=lowerCAmelCase, cls_token=lowerCAmelCase, pad_token=lowerCAmelCase, mask_token=lowerCAmelCase, tokenizer_file=lowerCAmelCase, src_lang=lowerCAmelCase, tgt_lang=lowerCAmelCase, additional_special_tokens=lowerCAmelCase, sp_model_kwargs=self.sp_model_kwargs, **lowerCAmelCase, ) lowercase : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCAmelCase ) ) lowercase : int = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token lowercase : Any = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab lowercase : Any = 1 lowercase : Dict = len(self.sp_model ) lowercase : Tuple = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase ) } lowercase : int = {v: k for k, v in self.lang_code_to_id.items()} lowercase : int = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) lowercase : List[str] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} lowercase : Optional[Any] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) lowercase : Tuple = src_lang if src_lang is not None else 'en_XX' lowercase : Union[str, Any] = self.lang_code_to_id[self._src_lang] lowercase : Dict = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : Optional[int] ) -> Union[str, Any]: lowercase : Optional[Any] = self.__dict__.copy() lowercase : Tuple = None lowercase : Union[str, Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Any, lowerCAmelCase : Dict ) -> Optional[int]: lowercase : Any = d # for backward compatibility if not hasattr(self, 'sp_model_kwargs' ): lowercase : List[str] = {} lowercase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def lowercase ( self : List[Any] ) -> Dict: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def lowercase ( self : Union[str, Any] ) -> str: return self._src_lang @src_lang.setter def lowercase ( self : Union[str, Any], lowerCAmelCase : str ) -> None: lowercase : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase ( self : str, lowerCAmelCase : List[int], lowerCAmelCase : Optional[List[int]] = None, lowerCAmelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase, token_ids_a=lowerCAmelCase, already_has_special_tokens=lowerCAmelCase ) lowercase : Optional[int] = [1] * len(self.prefix_tokens ) lowercase : Optional[int] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase )) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase )) + ([0] * len(lowerCAmelCase )) + suffix_ones def lowercase ( self : Any, lowerCAmelCase : List[int], lowerCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase ( self : Tuple, lowerCAmelCase : List[int], lowerCAmelCase : Optional[List[int]] = None ) -> List[int]: lowercase : Optional[Any] = [self.sep_token_id] lowercase : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase ( self : str, lowerCAmelCase : Optional[int], lowerCAmelCase : str, lowerCAmelCase : Optional[str], lowerCAmelCase : Optional[str], **lowerCAmelCase : str ) -> List[Any]: if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) lowercase : Optional[Any] = src_lang lowercase : int = self(lowerCAmelCase, add_special_tokens=lowerCAmelCase, return_tensors=lowerCAmelCase, **lowerCAmelCase ) lowercase : Any = self.convert_tokens_to_ids(lowerCAmelCase ) lowercase : Optional[int] = tgt_lang_id return inputs def lowercase ( self : List[Any] ) -> List[Any]: lowercase : List[str] = {self.convert_ids_to_tokens(lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase ( self : Union[str, Any], lowerCAmelCase : str ) -> List[str]: return self.sp_model.encode(lowerCAmelCase, out_type=lowerCAmelCase ) def lowercase ( self : Union[str, Any], lowerCAmelCase : List[Any] ) -> Union[str, Any]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowercase : Union[str, Any] = self.sp_model.PieceToId(lowerCAmelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowercase ( self : str, lowerCAmelCase : Tuple ) -> int: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowercase ( self : int, lowerCAmelCase : Tuple ) -> Optional[int]: lowercase : str = ''.join(lowerCAmelCase ).replace(lowerCAmelCase, ' ' ).strip() return out_string def lowercase ( self : List[Any], lowerCAmelCase : str, lowerCAmelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase : Union[str, Any] = os.path.join( lowerCAmelCase, (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file, lowerCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase, 'wb' ) as fi: lowercase : List[Any] = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase ) return (out_vocab_file,) def lowercase ( self : List[Any], lowerCAmelCase : List[str], lowerCAmelCase : str = "en_XX", lowerCAmelCase : Optional[List[str]] = None, lowerCAmelCase : str = "ro_RO", **lowerCAmelCase : Dict, ) -> BatchEncoding: lowercase : int = src_lang lowercase : List[Any] = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase, lowerCAmelCase, **lowerCAmelCase ) def lowercase ( self : List[Any] ) -> str: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase ( self : Any ) -> Dict: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase ( self : Optional[Any], lowerCAmelCase : Union[str, Any] ) -> None: lowercase : str = self.lang_code_to_id[src_lang] lowercase : Optional[int] = [] lowercase : Dict = [self.eos_token_id, self.cur_lang_code] def lowercase ( self : Optional[Any], lowerCAmelCase : str ) -> None: lowercase : List[str] = self.lang_code_to_id[lang] lowercase : Dict = [] lowercase : Dict = [self.eos_token_id, self.cur_lang_code]
53
"""simple docstring""" from collections.abc import Sequence def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase = False ) -> float: '''simple docstring''' if not arr: return 0 lowercase : List[str] = 0 if allow_empty_subarrays else float('-inf' ) lowercase : Dict = 0.0 for num in arr: lowercase : List[str] = max(0 if allow_empty_subarrays else num , curr_sum + num ) lowercase : List[Any] = max(_UpperCAmelCase , _UpperCAmelCase ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() _UpperCamelCase: Any = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(f'''{max_subarray_sum(nums) = }''')
53
1
'''simple docstring''' import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS logging.basicConfig(level=logging.DEBUG) __lowerCAmelCase = logging.getLogger() def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Dict: _a : Any = '\n'.join(lowerCAmelCase_ ) Path(lowerCAmelCase_ ).open('w' ).writelines(lowerCAmelCase_ ) __lowerCAmelCase = '''patrickvonplaten/t5-tiny-random''' __lowerCAmelCase = '''sshleifer/bart-tiny-random''' __lowerCAmelCase = '''sshleifer/tiny-mbart''' __lowerCAmelCase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) logging.disable(logging.CRITICAL) # remove noisy download output from tracebacks class __magic_name__ ( _UpperCamelCase ): def __lowercase ( self : List[Any] ,_UpperCAmelCase : Optional[int] ): _a : Tuple = Path(self.get_auto_remove_tmp_dir() ) / 'utest_input.source' _a : Any = input_file_name.parent / 'utest_output.txt' assert not output_file_name.exists() _a : Union[str, Any] = [' New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County.'] _dump_articles(_UpperCAmelCase ,_UpperCAmelCase ) _a : List[str] = str(Path(self.get_auto_remove_tmp_dir() ) / 'scores.json' ) _a : Union[str, Any] = 'translation_en_to_de' if model == T5_TINY else 'summarization' _a : Union[str, Any] = F""" run_eval_search.py {model} {input_file_name} {output_file_name} --score_path {score_path} --task {task} --num_beams 2 --length_penalty 2.0 """.split() with patch.object(_UpperCAmelCase ,'argv' ,_UpperCAmelCase ): run_generate() assert Path(_UpperCAmelCase ).exists() # os.remove(Path(output_file_name)) def __lowercase ( self : Optional[int] ): self.run_eval_tester(_UpperCAmelCase ) @parameterized.expand([BART_TINY, MBART_TINY] ) @slow def __lowercase ( self : List[Any] ,_UpperCAmelCase : Union[str, Any] ): self.run_eval_tester(_UpperCAmelCase ) @parameterized.expand([T5_TINY, MBART_TINY] ) @slow def __lowercase ( self : Union[str, Any] ,_UpperCAmelCase : Any ): _a : List[str] = Path(self.get_auto_remove_tmp_dir() ) / 'utest_input.source' _a : Tuple = input_file_name.parent / 'utest_output.txt' assert not output_file_name.exists() _a : Optional[int] = { 'en': ['Machine learning is great, isn\'t it?', 'I like to eat bananas', 'Tomorrow is another great day!'], 'de': [ 'Maschinelles Lernen ist großartig, oder?', 'Ich esse gerne Bananen', 'Morgen ist wieder ein toller Tag!', ], } _a : List[str] = Path(self.get_auto_remove_tmp_dir() ) _a : List[str] = str(tmp_dir / 'scores.json' ) _a : List[Any] = str(tmp_dir / 'val.target' ) _dump_articles(_UpperCAmelCase ,text['en'] ) _dump_articles(_UpperCAmelCase ,text['de'] ) _a : Optional[int] = 'translation_en_to_de' if model == T5_TINY else 'summarization' _a : str = F""" run_eval_search.py {model} {str(_UpperCAmelCase )} {str(_UpperCAmelCase )} --score_path {score_path} --reference_path {reference_path} --task {task} """.split() testargs.extend(['--search', 'num_beams=1:2 length_penalty=0.9:1.0'] ) with patch.object(_UpperCAmelCase ,'argv' ,_UpperCAmelCase ): with CaptureStdout() as cs: run_search() _a : List[Any] = [' num_beams | length_penalty', model, 'Best score args'] _a : int = ['Info'] if "translation" in task: expected_strings.append('bleu' ) else: expected_strings.extend(_UpperCAmelCase ) for w in expected_strings: assert w in cs.out for w in un_expected_strings: assert w not in cs.out assert Path(_UpperCAmelCase ).exists() os.remove(Path(_UpperCAmelCase ) )
89
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
13
0
import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class lowerCAmelCase_ : '''simple docstring''' def __init__( self : List[str] , _UpperCAmelCase : Tuple ): """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden UpperCAmelCase__ = deepcopy(_UpperCAmelCase ) elif os.path.exists(_UpperCAmelCase ): with io.open(_UpperCAmelCase , """r""" , encoding="""utf-8""" ) as f: UpperCAmelCase__ = json.load(_UpperCAmelCase ) else: try: UpperCAmelCase__ = baseaa.urlsafe_baadecode(_UpperCAmelCase ).decode("""utf-8""" ) UpperCAmelCase__ = json.loads(_UpperCAmelCase ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f'''Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}''' ) UpperCAmelCase__ = config self.set_stage_and_offload() def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" UpperCAmelCase__ = self.get_value("""zero_optimization.stage""" , -1 ) # offload UpperCAmelCase__ = False if self.is_zeroa() or self.is_zeroa(): UpperCAmelCase__ = set(["""cpu""", """nvme"""] ) UpperCAmelCase__ = set( [ self.get_value("""zero_optimization.offload_optimizer.device""" ), self.get_value("""zero_optimization.offload_param.device""" ), ] ) if len(offload_devices & offload_devices_valid ) > 0: UpperCAmelCase__ = True def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , _UpperCAmelCase : Optional[int] ): """simple docstring""" UpperCAmelCase__ = self.config # find the config node of interest if it exists UpperCAmelCase__ = ds_key_long.split(""".""" ) UpperCAmelCase__ = nodes.pop() for node in nodes: UpperCAmelCase__ = config.get(_UpperCAmelCase ) if config is None: return None, ds_key return config, ds_key def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[Any]=None ): """simple docstring""" UpperCAmelCase__ = self.find_config_node(_UpperCAmelCase ) if config is None: return default return config.get(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Tuple , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Tuple=False ): """simple docstring""" UpperCAmelCase__ = self.config # find the config node of interest if it exists UpperCAmelCase__ = ds_key_long.split(""".""" ) for node in nodes: UpperCAmelCase__ = config UpperCAmelCase__ = config.get(_UpperCAmelCase ) if config is None: if must_exist: raise ValueError(f'''Can\'t find {ds_key_long} entry in the config: {self.config}''' ) else: return # if found remove it if parent_config is not None: parent_config.pop(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , _UpperCAmelCase : Optional[int] ): """simple docstring""" UpperCAmelCase__ = self.get_value(_UpperCAmelCase ) return False if value is None else bool(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , _UpperCAmelCase : Dict ): """simple docstring""" UpperCAmelCase__ = self.get_value(_UpperCAmelCase ) return False if value is None else not bool(_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" return self._stage == 2 def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" return self._stage == 3 def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" return self._offload class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Optional[int] , _UpperCAmelCase : Dict ): """simple docstring""" UpperCAmelCase__ = engine def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , _UpperCAmelCase : Dict , **_UpperCAmelCase : str ): """simple docstring""" self.engine.backward(_UpperCAmelCase , **_UpperCAmelCase ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__( self : List[str] , _UpperCAmelCase : Dict ): """simple docstring""" super().__init__(_UpperCAmelCase , device_placement=_UpperCAmelCase , scaler=_UpperCAmelCase ) UpperCAmelCase__ = hasattr(self.optimizer , """overflow""" ) def SCREAMING_SNAKE_CASE__ ( self : int , _UpperCAmelCase : List[str]=None ): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" if self.__has_overflow__: return self.optimizer.overflow return False class lowerCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__( self : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : Tuple ): """simple docstring""" super().__init__(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Any , _UpperCAmelCase : str , _UpperCAmelCase : Optional[int]=0.001 , _UpperCAmelCase : List[str]=0 , **_UpperCAmelCase : List[str] ): """simple docstring""" UpperCAmelCase__ = params UpperCAmelCase__ = lr UpperCAmelCase__ = weight_decay UpperCAmelCase__ = kwargs class lowerCAmelCase_ : '''simple docstring''' def __init__( self : str , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : Union[str, Any]=0 , **_UpperCAmelCase : List[str] ): """simple docstring""" UpperCAmelCase__ = optimizer UpperCAmelCase__ = total_num_steps UpperCAmelCase__ = warmup_num_steps UpperCAmelCase__ = kwargs
359
'''simple docstring''' UpperCAmelCase_ = [sum(int(c, 1_0) ** 2 for c in i.__str__()) for i in range(1_0_0_0_0_0)] def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' UpperCAmelCase__ = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 100000] number //= 100000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution UpperCAmelCase_ = [None] * 1_0_0_0_0_0_0_0 UpperCAmelCase_ = True UpperCAmelCase_ = False def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore UpperCAmelCase__ = chain(next_number(SCREAMING_SNAKE_CASE__ ) ) UpperCAmelCase__ = number_chain while number < 10000000: UpperCAmelCase__ = number_chain number *= 10 return number_chain def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int = 10000000 ): '''simple docstring''' for i in range(1 , SCREAMING_SNAKE_CASE__ ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": import doctest doctest.testmod() print(f"{solution() = }")
61
0
from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def a__ ( __UpperCamelCase = True , *__UpperCamelCase , **__UpperCamelCase ): if not is_tqdm_available(): raise ImportError("Accelerate's `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`." ) SCREAMING_SNAKE_CASE_ = False if main_process_only: SCREAMING_SNAKE_CASE_ = PartialState().local_process_index == 0 return _tqdm(*__UpperCamelCase , **__UpperCamelCase , disable=__UpperCamelCase )
118
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A : str = logging.get_logger(__name__) def a__ ( __UpperCamelCase , __UpperCamelCase=False ): SCREAMING_SNAKE_CASE_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" SCREAMING_SNAKE_CASE_ = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def a__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: SCREAMING_SNAKE_CASE_ = "" else: SCREAMING_SNAKE_CASE_ = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) SCREAMING_SNAKE_CASE_ = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) SCREAMING_SNAKE_CASE_ = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE_ = in_proj_weight[ : config.hidden_size, : ] SCREAMING_SNAKE_CASE_ = in_proj_bias[: config.hidden_size] SCREAMING_SNAKE_CASE_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] SCREAMING_SNAKE_CASE_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] SCREAMING_SNAKE_CASE_ = in_proj_weight[ -config.hidden_size :, : ] SCREAMING_SNAKE_CASE_ = in_proj_bias[-config.hidden_size :] def a__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = dct.pop(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = val def a__ ( ): SCREAMING_SNAKE_CASE_ = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE_ = Image.open(requests.get(__UpperCamelCase , stream=__UpperCamelCase ).raw ) return im @torch.no_grad() def a__ ( __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = DeiTConfig() # all deit models have fine-tuned heads SCREAMING_SNAKE_CASE_ = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size SCREAMING_SNAKE_CASE_ = 1_0_0_0 SCREAMING_SNAKE_CASE_ = "huggingface/label-files" SCREAMING_SNAKE_CASE_ = "imagenet-1k-id2label.json" SCREAMING_SNAKE_CASE_ = json.load(open(hf_hub_download(__UpperCamelCase , __UpperCamelCase , repo_type="dataset" ) , "r" ) ) SCREAMING_SNAKE_CASE_ = {int(__UpperCamelCase ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_ = idalabel SCREAMING_SNAKE_CASE_ = {v: k for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_ = int(deit_name[-6:-4] ) SCREAMING_SNAKE_CASE_ = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): SCREAMING_SNAKE_CASE_ = 1_9_2 SCREAMING_SNAKE_CASE_ = 7_6_8 SCREAMING_SNAKE_CASE_ = 1_2 SCREAMING_SNAKE_CASE_ = 3 elif deit_name[9:].startswith("small" ): SCREAMING_SNAKE_CASE_ = 3_8_4 SCREAMING_SNAKE_CASE_ = 1_5_3_6 SCREAMING_SNAKE_CASE_ = 1_2 SCREAMING_SNAKE_CASE_ = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): SCREAMING_SNAKE_CASE_ = 1_0_2_4 SCREAMING_SNAKE_CASE_ = 4_0_9_6 SCREAMING_SNAKE_CASE_ = 2_4 SCREAMING_SNAKE_CASE_ = 1_6 # load original model from timm SCREAMING_SNAKE_CASE_ = timm.create_model(__UpperCamelCase , pretrained=__UpperCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys SCREAMING_SNAKE_CASE_ = timm_model.state_dict() SCREAMING_SNAKE_CASE_ = create_rename_keys(__UpperCamelCase , __UpperCamelCase ) for src, dest in rename_keys: rename_key(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) read_in_q_k_v(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # load HuggingFace model SCREAMING_SNAKE_CASE_ = DeiTForImageClassificationWithTeacher(__UpperCamelCase ).eval() model.load_state_dict(__UpperCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor SCREAMING_SNAKE_CASE_ = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 SCREAMING_SNAKE_CASE_ = DeiTImageProcessor(size=__UpperCamelCase , crop_size=config.image_size ) SCREAMING_SNAKE_CASE_ = image_processor(images=prepare_img() , return_tensors="pt" ) SCREAMING_SNAKE_CASE_ = encoding["pixel_values"] SCREAMING_SNAKE_CASE_ = model(__UpperCamelCase ) SCREAMING_SNAKE_CASE_ = timm_model(__UpperCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__UpperCamelCase , outputs.logits , atol=1E-3 ) Path(__UpperCamelCase ).mkdir(exist_ok=__UpperCamelCase ) print(F'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__UpperCamelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": A : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--deit_name", default="vit_deit_base_distilled_patch16_224", type=str, help="Name of the DeiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) A : Dict = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
118
1
"""simple docstring""" import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class _UpperCAmelCase ( __snake_case ): '''simple docstring''' def __init__(self , *a_ , a_=None , a_=None , **a_ ): '''simple docstring''' super().__init__(*a_ , **a_ ) __snake_case : List[str] = eval_examples __snake_case : Optional[Any] = post_process_function def SCREAMING_SNAKE_CASE (self , a_=None , a_=None , a_=None , a_ = "eval" ): '''simple docstring''' __snake_case : Union[str, Any] = self.eval_dataset if eval_dataset is None else eval_dataset __snake_case : Union[str, Any] = self.get_eval_dataloader(a_ ) __snake_case : List[str] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. __snake_case : Optional[Any] = self.compute_metrics __snake_case : Union[str, Any] = None __snake_case : str = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop __snake_case : str = time.time() try: __snake_case : Dict = eval_loop( a_ , description='''Evaluation''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=a_ , metric_key_prefix=a_ , ) finally: __snake_case : str = compute_metrics __snake_case : Optional[Any] = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( a_ , a_ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default __snake_case : Optional[Any] = self.post_process_function(a_ , a_ , output.predictions ) __snake_case : Any = self.compute_metrics(a_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): __snake_case : List[str] = metrics.pop(a_ ) metrics.update(output.metrics ) else: __snake_case : str = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(a_ ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) __snake_case : str = self.callback_handler.on_evaluate(self.args , self.state , self.control , a_ ) return metrics def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_=None , a_ = "test" ): '''simple docstring''' __snake_case : Any = self.get_test_dataloader(a_ ) # Temporarily disable metric computation, we will do it in the loop here. __snake_case : List[str] = self.compute_metrics __snake_case : List[Any] = None __snake_case : List[Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop __snake_case : Dict = time.time() try: __snake_case : List[Any] = eval_loop( a_ , description='''Prediction''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=a_ , metric_key_prefix=a_ , ) finally: __snake_case : Optional[int] = compute_metrics __snake_case : str = self.args.eval_batch_size * self.args.world_size if f"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[f"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( a_ , a_ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output __snake_case : Optional[int] = self.post_process_function(a_ , a_ , output.predictions , '''predict''' ) __snake_case : Union[str, Any] = self.compute_metrics(a_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f"""{metric_key_prefix}_""" ): __snake_case : int = metrics.pop(a_ ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=a_ )
350
"""simple docstring""" def lowercase ( ) ->int: """simple docstring""" return [ a * b * (1_000 - a - b) for a in range(1 , 999 ) for b in range(_snake_case , 999 ) if (a * a + b * b == (1_000 - a - b) ** 2) ][0] if __name__ == "__main__": print(F'{solution() = }')
24
0
def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : int = 2_000_000) -> int: '''simple docstring''' __UpperCamelCase : Any = [0 for i in range(n + 1)] __UpperCamelCase : int = 1 __UpperCamelCase : List[str] = 1 for i in range(2 , int(n**0.5) + 1): if primality_list[i] == 0: for j in range(i * i , n + 1 , _lowerCamelCase): __UpperCamelCase : int = 1 __UpperCamelCase : str = 0 for i in range(_lowerCamelCase): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(f"{solution() = }")
232
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : List[Any] = logging.get_logger(__name__) lowercase : Any = { 'google/pegasus-large': 'https://huggingface.co/google/pegasus-large/resolve/main/config.json', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class lowerCamelCase__ ( __lowercase): '''simple docstring''' _A = 'pegasus' _A = ['past_key_values'] _A = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self :Dict , a :Dict=5_0_2_6_5 , a :Dict=1_0_2_4 , a :Union[str, Any]=1_2 , a :Any=4_0_9_6 , a :str=1_6 , a :str=1_2 , a :Optional[Any]=4_0_9_6 , a :int=1_6 , a :Optional[int]=0.0 , a :Optional[int]=0.0 , a :List[Any]=True , a :Union[str, Any]=True , a :int="gelu" , a :Dict=1_0_2_4 , a :List[Any]=0.1 , a :List[str]=0.0 , a :List[Any]=0.0 , a :str=0.02 , a :int=0 , a :Any=False , a :Dict=0 , a :int=1 , a :Optional[Any]=1 , **a :Optional[int] , ) -> str: __UpperCamelCase : List[Any] = vocab_size __UpperCamelCase : Union[str, Any] = max_position_embeddings __UpperCamelCase : str = d_model __UpperCamelCase : Dict = encoder_ffn_dim __UpperCamelCase : int = encoder_layers __UpperCamelCase : int = encoder_attention_heads __UpperCamelCase : List[Any] = decoder_ffn_dim __UpperCamelCase : List[Any] = decoder_layers __UpperCamelCase : List[str] = decoder_attention_heads __UpperCamelCase : str = dropout __UpperCamelCase : Union[str, Any] = attention_dropout __UpperCamelCase : List[str] = activation_dropout __UpperCamelCase : Optional[Any] = activation_function __UpperCamelCase : Tuple = init_std __UpperCamelCase : Optional[int] = encoder_layerdrop __UpperCamelCase : Union[str, Any] = decoder_layerdrop __UpperCamelCase : Optional[Any] = use_cache __UpperCamelCase : Union[str, Any] = encoder_layers __UpperCamelCase : int = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=a , eos_token_id=a , is_encoder_decoder=a , decoder_start_token_id=a , forced_eos_token_id=a , **a , ) @property def _lowerCamelCase ( self :Dict ) -> int: return self.encoder_attention_heads @property def _lowerCamelCase ( self :Optional[Any] ) -> int: return self.d_model
232
1
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowercase_ ( ) -> Dict: """simple docstring""" snake_case = ArgumentParser("Transformers CLI tool" , usage="transformers-cli <command> [<args>]" ) snake_case = parser.add_subparsers(help="transformers-cli command helpers" ) # Register commands ConvertCommand.register_subcommand(_lowerCAmelCase ) DownloadCommand.register_subcommand(_lowerCAmelCase ) EnvironmentCommand.register_subcommand(_lowerCAmelCase ) RunCommand.register_subcommand(_lowerCAmelCase ) ServeCommand.register_subcommand(_lowerCAmelCase ) UserCommands.register_subcommand(_lowerCAmelCase ) AddNewModelCommand.register_subcommand(_lowerCAmelCase ) AddNewModelLikeCommand.register_subcommand(_lowerCAmelCase ) LfsCommands.register_subcommand(_lowerCAmelCase ) PTtoTFCommand.register_subcommand(_lowerCAmelCase ) # Let's go snake_case = parser.parse_args() if not hasattr(_lowerCAmelCase , "func" ): parser.print_help() exit(1 ) # Run snake_case = args.func(_lowerCAmelCase ) service.run() if __name__ == "__main__": main()
354
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy _A = logging.getLogger(__name__) _A = "pytorch_model.bin" @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co."} , ) @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the training data."} ) UpperCAmelCase__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the data to predict on."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "A csv or a json file containing the validation data."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default=A_ , metadata={"help": "The name of the task to train on."} , ) UpperCAmelCase__ : Optional[List[str]] = dataclasses.field( default=A_ , metadata={"help": "The list of labels for the task."} ) @dataclasses.dataclass class lowerCamelCase : UpperCAmelCase__ : str = dataclasses.field( metadata={"help": "The output directory where the model predictions and checkpoints will be written."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default="accuracy" , metadata={"help": "The evaluation metric used for the task."} ) UpperCAmelCase__ : Optional[str] = dataclasses.field( default="no" , metadata={ "help": "The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]" } , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=10 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Optional[float] = dataclasses.field( default=0.0 , metadata={ "help": "How much the specified evaluation metric must improve to satisfy early stopping conditions." } , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to filter the pseudo-labeled data based on the confidence score."} , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to filter the pseudo-labeled data based on the validation performance."} , ) UpperCAmelCase__ : Optional[bool] = dataclasses.field( default=A_ , metadata={"help": "Whether to fine-tune on labeled data after pseudo training."} , ) UpperCAmelCase__ : Optional[float] = dataclasses.field( default=0.0 , metadata={"help": "Confidence threshold for pseudo-labeled data filtering."} , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=1_00 , metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} , ) UpperCAmelCase__ : Optional[int] = dataclasses.field( default=A_ , metadata={"help": "Random seed for initialization."} , ) def lowercase_ ( A__ , A__ , A__ , A__ , A__ , A__ ) -> Union[str, Any]: """simple docstring""" snake_case = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: snake_case = dataset.filter(lambda A__ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 snake_case = int(eval_result * len(A__ ) ) print(A__ ) snake_case = dataset.sort("probability" , reverse=A__ ) snake_case = dataset.select(range(A__ ) ) snake_case = dataset.remove_columns(["label", "probability"] ) snake_case = dataset.rename_column("prediction" , "label" ) snake_case = dataset.map(lambda A__ : {"label": idalabel[example["label"]]} ) snake_case = dataset.shuffle(seed=args.seed ) snake_case = os.path.join(A__ , F'train_pseudo.{args.data_file_extension}' ) if args.data_file_extension == "csv": dataset.to_csv(A__ , index=A__ ) else: dataset.to_json(A__ ) def lowercase_ ( A__ , A__ , A__ , A__ , **A__ ) -> List[Any]: """simple docstring""" snake_case = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() snake_case = STModelArguments(model_name_or_path=A__ ) snake_case = STDataArguments(train_file=A__ , infer_file=A__ ) snake_case = STTrainingArguments(output_dir=A__ ) snake_case = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(A__ ).items(): setattr(A__ , A__ , A__ ) for key, value in kwargs.items(): if hasattr(A__ , A__ ): setattr(A__ , A__ , A__ ) # Sanity checks snake_case = {} snake_case = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None snake_case = args.train_file snake_case = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None snake_case = args.eval_file for key in data_files: snake_case = data_files[key].split("." )[-1] assert extension in ["csv", "json"], F'`{key}_file` should be a csv or a json file.' if args.data_file_extension is None: snake_case = extension else: assert extension == args.data_file_extension, F'`{key}_file` should be a {args.data_file_extension} file`.' assert ( args.eval_metric in datasets.list_metrics() ), F'{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.' # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("Creating the initial data directory for self-training..." ) snake_case = F'{args.output_dir}/self-train_iter-{{}}'.format snake_case = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=A__ ) os.makedirs(A__ , exist_ok=A__ ) accelerator.wait_for_everyone() snake_case = None snake_case = None snake_case = 0 snake_case = False # Show the progress bar snake_case = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): snake_case = data_dir_format(A__ ) assert os.path.exists(A__ ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 snake_case = os.path.join(A__ , "stage-1" ) snake_case = { "accelerator": accelerator, "model_name_or_path": args.model_name_or_path, "cache_dir": args.cache_dir, "do_train": True, "train_file": data_files["train"] if iteration == 0 else data_files["train_pseudo"], "do_eval": True if args.eval_file is not None else False, "eval_file": data_files["eval"], "do_predict": True, "infer_file": data_files["infer"], "task_name": args.task_name, "label_list": args.label_list, "output_dir": current_output_dir, "eval_metric": args.eval_metric, "evaluation_strategy": args.evaluation_strategy, "early_stopping_patience": args.early_stopping_patience, "early_stopping_threshold": args.early_stopping_threshold, "seed": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(A__ , A__ ): arguments_dict.update({key: value} ) snake_case = os.path.join(A__ , "best-checkpoint" , A__ ) if os.path.exists(A__ ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1." , A__ , A__ , ) else: logger.info("***** Running self-training: iteration: %d, stage: 1 *****" , A__ ) finetune(**A__ ) accelerator.wait_for_everyone() assert os.path.exists(A__ ) logger.info("Self-training job completed: iteration: %d, stage: 1." , A__ ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data snake_case = os.path.join(A__ , "best-checkpoint" ) snake_case = os.path.join(A__ , "stage-2" ) # Update arguments_dict snake_case = model_path snake_case = data_files["train"] snake_case = current_output_dir snake_case = os.path.join(A__ , "best-checkpoint" , A__ ) if os.path.exists(A__ ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2." , A__ , A__ , ) else: logger.info("***** Running self-training: iteration: %d, stage: 2 *****" , A__ ) finetune(**A__ ) accelerator.wait_for_everyone() assert os.path.exists(A__ ) logger.info("Self-training job completed: iteration: %d, stage: 2." , A__ ) snake_case = iteration snake_case = data_dir_format(iteration + 1 ) snake_case = AutoConfig.from_pretrained(os.path.join(A__ , "best-checkpoint" ) ) snake_case = config.idalabel snake_case = os.path.join(A__ , "eval_results_best-checkpoint.json" ) snake_case = os.path.join(A__ , "test_results_best-checkpoint.json" ) assert os.path.exists(A__ ) with open(A__ , "r" ) as f: snake_case = float(json.load(A__ )[args.eval_metric] ) snake_case = os.path.join(A__ , "infer_output_best-checkpoint.csv" ) assert os.path.exists(A__ ) # Loading the dataset from local csv or json files. snake_case = load_dataset(args.data_file_extension , data_files={"data": data_files["infer"]} )["data"] snake_case = load_dataset("csv" , data_files={"data": infer_output_file} )["data"] if accelerator.is_main_process: os.makedirs(A__ , exist_ok=A__ ) shutil.copy(A__ , os.path.join(A__ , F'eval_results_iter-{iteration}.json' ) ) if os.path.exists(A__ ): shutil.copy(A__ , os.path.join(A__ , F'test_results_iter-{iteration}.json' ) ) create_pseudo_labeled_data(A__ , A__ , A__ , A__ , A__ , A__ ) accelerator.wait_for_everyone() snake_case = os.path.join(A__ , F'train_pseudo.{args.data_file_extension}' ) if args.evaluation_strategy != IntervalStrategy.NO.value: snake_case = eval_result if best_iteration is None: snake_case = new_iteration snake_case = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: snake_case = new_iteration snake_case = new_eval_result snake_case = 0 else: if new_eval_result == best_eval_result: snake_case = new_iteration snake_case = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: snake_case = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("Best iteration: %d" , A__ ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , A__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A__ , F'eval_results_iter-{iteration}.json' ) , os.path.join(A__ , "eval_results_best-iteration.json" ) , ) else: # Assume that the last iteration is the best logger.info("Best iteration: %d" , args.max_selftrain_iterations - 1 ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , A__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(A__ , F'eval_results_iter-{args.max_selftrain_iterations - 1}.json' ) , os.path.join(A__ , "eval_results_best-iteration.json" ) , )
137
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a__ : Tuple =logging.get_logger(__name__) a__ : int ={ '''microsoft/beit-base-patch16-224-pt22k''': ( '''https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json''' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="beit" def __init__( self : List[Any] , __A : List[Any]=8_1_9_2 , __A : int=7_6_8 , __A : Tuple=1_2 , __A : Optional[Any]=1_2 , __A : Union[str, Any]=3_0_7_2 , __A : Optional[Any]="gelu" , __A : Tuple=0.0 , __A : int=0.0 , __A : Optional[int]=0.02 , __A : Tuple=1e-12 , __A : Union[str, Any]=2_2_4 , __A : Tuple=1_6 , __A : Any=3 , __A : List[Any]=False , __A : str=False , __A : Any=False , __A : Optional[Any]=False , __A : Optional[Any]=0.1 , __A : Optional[int]=0.1 , __A : Optional[Any]=True , __A : Any=[3, 5, 7, 1_1] , __A : str=[1, 2, 3, 6] , __A : List[str]=True , __A : Union[str, Any]=0.4 , __A : Dict=2_5_6 , __A : Any=1 , __A : List[str]=False , __A : Tuple=2_5_5 , **__A : List[str] , ): super().__init__(**__A ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_act __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = initializer_range __UpperCamelCase = layer_norm_eps __UpperCamelCase = image_size __UpperCamelCase = patch_size __UpperCamelCase = num_channels __UpperCamelCase = use_mask_token __UpperCamelCase = use_absolute_position_embeddings __UpperCamelCase = use_relative_position_bias __UpperCamelCase = use_shared_relative_position_bias __UpperCamelCase = layer_scale_init_value __UpperCamelCase = drop_path_rate __UpperCamelCase = use_mean_pooling # decode head attributes (semantic segmentation) __UpperCamelCase = out_indices __UpperCamelCase = pool_scales # auxiliary head attributes (semantic segmentation) __UpperCamelCase = use_auxiliary_head __UpperCamelCase = auxiliary_loss_weight __UpperCamelCase = auxiliary_channels __UpperCamelCase = auxiliary_num_convs __UpperCamelCase = auxiliary_concat_input __UpperCamelCase = semantic_loss_ignore_index class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any =version.parse("1.11" ) @property def _lowerCamelCase ( self : Union[str, Any] ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def _lowerCamelCase ( self : Tuple ): return 1e-4
53
'''simple docstring''' import os import numpy import onnx def lowercase__ ( __lowercase : Optional[int] , __lowercase : Union[str, Any] ) -> Dict: """simple docstring""" __UpperCamelCase = a.name __UpperCamelCase = b.name __UpperCamelCase = '' __UpperCamelCase = '' __UpperCamelCase = a == b __UpperCamelCase = name_a __UpperCamelCase = name_b return res def lowercase__ ( __lowercase : int , __lowercase : int , __lowercase : List[Any] ) -> Optional[int]: """simple docstring""" for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(__lowercase , __lowercase ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) _graph_replace_input_with(node_proto.attribute[1].g , __lowercase , __lowercase ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , __lowercase , __lowercase ) def lowercase__ ( __lowercase : int , __lowercase : List[Any] , __lowercase : Dict ) -> int: """simple docstring""" for n in graph_proto.node: _node_replace_input_with(__lowercase , __lowercase , __lowercase ) def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any] , __lowercase : str ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i __UpperCamelCase = inits[i].name __UpperCamelCase = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , __lowercase , __lowercase ) def lowercase__ ( __lowercase : Dict ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = os.path.dirname(__lowercase ) __UpperCamelCase = os.path.basename(__lowercase ) __UpperCamelCase = onnx.load(os.path.join(__lowercase , __lowercase ) ) __UpperCamelCase = list(model.graph.initializer ) __UpperCamelCase = set() __UpperCamelCase = {} __UpperCamelCase = [] __UpperCamelCase = 0 for i in range(len(__lowercase ) ): if i in dup_set: continue for j in range(i + 1 , len(__lowercase ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(__lowercase ) dup_set.add(__lowercase ) __UpperCamelCase = inits[j].data_type __UpperCamelCase = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , __lowercase ) total_reduced_size += mem_size __UpperCamelCase = inits[i].name __UpperCamelCase = inits[j].name if name_i in dup_map: dup_map[name_i].append(__lowercase ) else: __UpperCamelCase = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) __UpperCamelCase = sorted(__lowercase ) _remove_dup_initializers_from_model(__lowercase , __lowercase , __lowercase ) __UpperCamelCase = 'optimized_' + model_file_name __UpperCamelCase = os.path.join(__lowercase , __lowercase ) onnx.save(__lowercase , __lowercase ) return new_model
53
1
import collections import inspect import unittest from typing import Dict, List, Tuple from transformers import MaskFormerSwinConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, torch_device from transformers.utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MaskFormerSwinBackbone from transformers.models.maskformer import MaskFormerSwinModel class SCREAMING_SNAKE_CASE_ : def __init__( self : Optional[Any] , _A : Optional[Any] , _A : Tuple=13 , _A : Any=32 , _A : Tuple=2 , _A : List[Any]=3 , _A : Any=16 , _A : Dict=[1, 2, 1] , _A : str=[2, 2, 4] , _A : List[str]=2 , _A : Union[str, Any]=2.0 , _A : Union[str, Any]=True , _A : int=0.0 , _A : List[Any]=0.0 , _A : int=0.1 , _A : Any="gelu" , _A : Optional[int]=False , _A : Optional[Any]=True , _A : Tuple=0.0_2 , _A : Tuple=1E-5 , _A : int=True , _A : Any=None , _A : str=True , _A : Tuple=10 , _A : str=8 , _A : List[str]=["stage1", "stage2", "stage3"] , _A : List[str]=[1, 2, 3] , ) -> List[str]: """simple docstring""" snake_case_ : Optional[int] = parent snake_case_ : int = batch_size snake_case_ : Optional[Any] = image_size snake_case_ : Any = patch_size snake_case_ : List[str] = num_channels snake_case_ : int = embed_dim snake_case_ : Optional[int] = depths snake_case_ : Union[str, Any] = num_heads snake_case_ : int = window_size snake_case_ : Optional[int] = mlp_ratio snake_case_ : Union[str, Any] = qkv_bias snake_case_ : Optional[int] = hidden_dropout_prob snake_case_ : Optional[int] = attention_probs_dropout_prob snake_case_ : Dict = drop_path_rate snake_case_ : List[str] = hidden_act snake_case_ : Tuple = use_absolute_embeddings snake_case_ : str = patch_norm snake_case_ : int = layer_norm_eps snake_case_ : Union[str, Any] = initializer_range snake_case_ : Tuple = is_training snake_case_ : Union[str, Any] = scope snake_case_ : Union[str, Any] = use_labels snake_case_ : Union[str, Any] = type_sequence_label_size snake_case_ : List[str] = encoder_stride snake_case_ : List[Any] = out_features snake_case_ : int = out_indices def UpperCAmelCase_ ( self : Any ) -> Dict: """simple docstring""" snake_case_ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ : Dict = None if self.use_labels: snake_case_ : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ : Optional[int] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self : List[Any] ) -> Tuple: """simple docstring""" return MaskFormerSwinConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCAmelCase_ ( self : Optional[int] , _A : Optional[Any] , _A : int , _A : Tuple ) -> Union[str, Any]: """simple docstring""" snake_case_ : Dict = MaskFormerSwinModel(config=_A ) model.to(_A ) model.eval() snake_case_ : Any = model(_A ) snake_case_ : Tuple = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ : Optional[Any] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCAmelCase_ ( self : Union[str, Any] , _A : List[str] , _A : int , _A : Any ) -> Dict: """simple docstring""" snake_case_ : Union[str, Any] = MaskFormerSwinBackbone(config=_A ) model.to(_A ) model.eval() snake_case_ : str = model(_A ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [13, 16, 16, 16] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , [16, 32, 64] ) # verify ValueError with self.parent.assertRaises(_A ): snake_case_ : int = ['stem'] snake_case_ : List[str] = MaskFormerSwinBackbone(config=_A ) def UpperCAmelCase_ ( self : Tuple ) -> str: """simple docstring""" snake_case_ : Dict = self.prepare_config_and_inputs() snake_case_ ,snake_case_ ,snake_case_ : Union[str, Any] = config_and_inputs snake_case_ : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE_ ( snake_case_ , snake_case_ , unittest.TestCase ): __magic_name__: str = ( ( MaskFormerSwinModel, MaskFormerSwinBackbone, ) if is_torch_available() else () ) __magic_name__: List[str] = {"feature-extraction": MaskFormerSwinModel} if is_torch_available() else {} __magic_name__: Optional[Any] = False __magic_name__: List[str] = False __magic_name__: Tuple = False __magic_name__: int = False __magic_name__: Optional[int] = False def UpperCAmelCase_ ( self : Union[str, Any] ) -> int: """simple docstring""" snake_case_ : List[str] = MaskFormerSwinModelTester(self ) snake_case_ : str = ConfigTester(self , config_class=_A , embed_dim=37 ) @require_torch_multi_gpu @unittest.skip( reason=( '`MaskFormerSwinModel` outputs `hidden_states_spatial_dimensions` which doesn\'t work well with' ' `nn.DataParallel`' ) ) def UpperCAmelCase_ ( self : str ) -> Dict: """simple docstring""" pass def UpperCAmelCase_ ( self : List[Any] ) -> Dict: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: """simple docstring""" return def UpperCAmelCase_ ( self : Any ) -> str: """simple docstring""" snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Tuple: """simple docstring""" snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_A ) @unittest.skip('Swin does not use inputs_embeds' ) def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip('Swin does not support feedforward chunking' ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> str: """simple docstring""" pass def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: """simple docstring""" snake_case_ ,snake_case_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : str = model_class(_A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_A , nn.Linear ) ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Tuple: """simple docstring""" snake_case_ ,snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ : str = model_class(_A ) snake_case_ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ : int = [*signature.parameters.keys()] snake_case_ : int = ['pixel_values'] self.assertListEqual(arg_names[:1] , _A ) @unittest.skip(reason='MaskFormerSwin is only used as backbone and doesn\'t support output_attentions' ) def UpperCAmelCase_ ( self : Dict ) -> str: """simple docstring""" pass @unittest.skip(reason='MaskFormerSwin is only used as an internal backbone' ) def UpperCAmelCase_ ( self : List[str] ) -> str: """simple docstring""" pass def UpperCAmelCase_ ( self : Tuple , _A : Optional[Any] , _A : int , _A : Tuple , _A : List[str] ) -> str: """simple docstring""" snake_case_ : List[str] = model_class(_A ) model.to(_A ) model.eval() with torch.no_grad(): snake_case_ : Any = model(**self._prepare_for_class(_A , _A ) ) snake_case_ : str = outputs.hidden_states snake_case_ : List[str] = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_A ) , _A ) # Swin has a different seq_length snake_case_ : Union[str, Any] = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ : List[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" snake_case_ ,snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Any = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case_ : Tuple = True self.check_hidden_states_output(_A , _A , _A , _A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ : List[Any] = True self.check_hidden_states_output(_A , _A , _A , _A ) def UpperCAmelCase_ ( self : str ) -> str: """simple docstring""" snake_case_ ,snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : List[Any] = 3 snake_case_ : Optional[int] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ : Optional[Any] = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ : Tuple = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ : Tuple = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case_ : Dict = True self.check_hidden_states_output(_A , _A , _A , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ : Tuple = True self.check_hidden_states_output(_A , _A , _A , (padded_height, padded_width) ) @unittest.skip(reason='MaskFormerSwin doesn\'t have pretrained checkpoints' ) def UpperCAmelCase_ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='This will be fixed once MaskFormerSwin is replaced by native Swin' ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: """simple docstring""" pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" snake_case_ ,snake_case_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() def set_nan_tensor_to_zero(_A : Tuple ): snake_case_ : List[Any] = 0 return t def check_equivalence(_A : Tuple , _A : int , _A : Tuple , _A : List[str]={} ): with torch.no_grad(): snake_case_ : str = model(**_A , return_dict=_A , **_A ) snake_case_ : Tuple = model(**_A , return_dict=_A , **_A ).to_tuple() def recursive_check(_A : Optional[Any] , _A : str ): if isinstance(_A , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_A , _A ): recursive_check(_A , _A ) elif isinstance(_A , _A ): for tuple_iterable_value, dict_iterable_value in zip( tuple_object.values() , dict_object.values() ): recursive_check(_A , _A ) elif tuple_object is None: return else: self.assertTrue( torch.allclose( set_nan_tensor_to_zero(_A ) , set_nan_tensor_to_zero(_A ) , atol=1E-5 ) , msg=( 'Tuple and dict output are not equal. Difference:' F""" {torch.max(torch.abs(tuple_object - dict_object ) )}. Tuple has `nan`:""" F""" {torch.isnan(_A ).any()} and `inf`: {torch.isinf(_A )}. Dict has""" F""" `nan`: {torch.isnan(_A ).any()} and `inf`: {torch.isinf(_A )}.""" ) , ) recursive_check(_A , _A ) for model_class in self.all_model_classes: snake_case_ : Any = model_class(_A ) model.to(_A ) model.eval() snake_case_ : Dict = self._prepare_for_class(_A , _A ) snake_case_ : List[str] = self._prepare_for_class(_A , _A ) check_equivalence(_A , _A , _A ) snake_case_ : Union[str, Any] = self._prepare_for_class(_A , _A , return_labels=_A ) snake_case_ : Any = self._prepare_for_class(_A , _A , return_labels=_A ) check_equivalence(_A , _A , _A ) snake_case_ : str = self._prepare_for_class(_A , _A ) snake_case_ : Union[str, Any] = self._prepare_for_class(_A , _A ) check_equivalence(_A , _A , _A , {'output_hidden_states': True} ) snake_case_ : int = self._prepare_for_class(_A , _A , return_labels=_A ) snake_case_ : Optional[Any] = self._prepare_for_class(_A , _A , return_labels=_A ) check_equivalence(_A , _A , _A , {'output_hidden_states': True} ) @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase , snake_case_ ): __magic_name__: List[str] = (MaskFormerSwinBackbone,) if is_torch_available() else () __magic_name__: int = MaskFormerSwinConfig def UpperCAmelCase_ ( self : int ) -> Tuple: """simple docstring""" snake_case_ : Optional[Any] = MaskFormerSwinModelTester(self ) def UpperCAmelCase_ ( self : Any ) -> Optional[Any]: """simple docstring""" snake_case_ ,snake_case_ : int = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ : Any = inputs_dict['pixel_values'].shape[0] for backbone_class in self.all_model_classes: snake_case_ : str = backbone_class(_A ) backbone.to(_A ) backbone.eval() snake_case_ : List[str] = backbone(**_A ) # Test default outputs and verify feature maps self.assertIsInstance(outputs.feature_maps , _A ) self.assertTrue(len(outputs.feature_maps ) == len(backbone.channels ) ) for feature_map, n_channels in zip(outputs.feature_maps , backbone.channels ): self.assertTrue(feature_map.shape[:2] , (batch_size, n_channels) ) self.assertIsNone(outputs.hidden_states ) self.assertIsNone(outputs.attentions ) # Test output_hidden_states=True snake_case_ : Any = backbone(**_A , output_hidden_states=_A ) self.assertIsNotNone(outputs.hidden_states ) self.assertTrue(len(outputs.hidden_states ) , len(backbone.stage_names ) ) # We skip the stem layer for hidden_states, n_channels in zip(outputs.hidden_states[1:] , backbone.channels ): for hidden_state in hidden_states: # Hidden states are in the format (batch_size, (height * width), n_channels) snake_case_ ,snake_case_ ,snake_case_ : Any = hidden_state.shape self.assertTrue((h_batch_size, h_n_channels) , (batch_size, n_channels) ) # Test output_attentions=True if self.has_attentions: snake_case_ : Any = backbone(**_A , output_attentions=_A ) self.assertIsNotNone(outputs.attentions )
88
def SCREAMING_SNAKE_CASE__ ( __a , __a = False ): if not isinstance(__a , __a ): snake_case_ : str = f"""Expected string as input, found {type(__a )}""" raise ValueError(__a ) if not isinstance(__a , __a ): snake_case_ : int = f"""Expected boolean as use_pascal parameter, found {type(__a )}""" raise ValueError(__a ) snake_case_ : Union[str, Any] = input_str.split('_' ) snake_case_ : int = 0 if use_pascal else 1 snake_case_ : List[Any] = words[start_index:] snake_case_ : str = [word[0].upper() + word[1:] for word in words_to_capitalize] snake_case_ : Optional[Any] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
88
1
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def A () -> Dict: """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(__A ): requests.request('''GET''' , '''https://huggingface.co''' ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request('''GET''' , '''https://huggingface.co''' , timeout=1.0 ) @pytest.mark.integration def A () -> List[Any]: """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request('''GET''' , '''https://huggingface.co''' ) def A () -> List[Any]: """simple docstring""" with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(__A ): http_head('''https://huggingface.co''' )
51
"""simple docstring""" import argparse from collections import defaultdict def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = f"""{file}_{class_name}_{test_name}""" done_test[_id] += 1 with open(__lowerCamelCase, "r" ) as f: UpperCAmelCase_ : List[Any] = f.readlines() UpperCAmelCase_ : int = f"""class {class_name}(""" UpperCAmelCase_ : Optional[Any] = f"""{4 * " "}def {test_name}(""" UpperCAmelCase_ : Optional[Any] = f"""{8 * " "}{correct_line.split()[0]}""" UpperCAmelCase_ : Tuple = f"""{16 * " "}{correct_line.split()[0]}""" UpperCAmelCase_ : int = False UpperCAmelCase_ : Union[str, Any] = False UpperCAmelCase_ : str = False UpperCAmelCase_ : Optional[Any] = False UpperCAmelCase_ : List[str] = 0 UpperCAmelCase_ : Optional[int] = 0 UpperCAmelCase_ : int = [] for line in lines: if line.startswith(__lowerCamelCase ): UpperCAmelCase_ : Tuple = True elif in_class and line.startswith(__lowerCamelCase ): UpperCAmelCase_ : Optional[int] = True elif in_class and in_func and (line.startswith(__lowerCamelCase ) or line.startswith(__lowerCamelCase )): UpperCAmelCase_ : Any = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: UpperCAmelCase_ : Union[str, Any] = True if in_class and in_func and in_line: if ")" not in line: continue else: UpperCAmelCase_ : Any = True if in_class and in_func and in_line and insert_line: new_lines.append(f"""{spaces * " "}{correct_line}""" ) UpperCAmelCase_ : int = False else: new_lines.append(__lowerCamelCase ) with open(__lowerCamelCase, "w" ) as f: for line in new_lines: f.write(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase=None ): if fail is not None: with open(__lowerCamelCase, "r" ) as f: UpperCAmelCase_ : Tuple = {l.strip() for l in f.readlines()} else: UpperCAmelCase_ : str = None with open(__lowerCamelCase, "r" ) as f: UpperCAmelCase_ : Optional[int] = f.readlines() UpperCAmelCase_ : Any = defaultdict(__lowerCamelCase ) for line in correct_lines: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Any = line.split(";" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument('--correct_filename', help='filename of tests with expected result') parser.add_argument('--fail_filename', help='filename of test failures', type=str, default=None) _a = parser.parse_args() main(args.correct_filename, args.fail_filename)
61
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: UpperCamelCase_ = None UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} UpperCamelCase_ = { 'vocab_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model', }, 'tokenizer_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json', }, } UpperCamelCase_ = { 'xlnet-base-cased': None, 'xlnet-large-cased': None, } UpperCamelCase_ = '▁' # Segments (not really needed) UpperCamelCase_ = 0 UpperCamelCase_ = 1 UpperCamelCase_ = 2 UpperCamelCase_ = 3 UpperCamelCase_ = 4 class snake_case ( SCREAMING_SNAKE_CASE_ ): a_ : Optional[Any] = VOCAB_FILES_NAMES a_ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP a_ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ : List[str] = """left""" a_ : Optional[Any] = XLNetTokenizer def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase=False , __UpperCAmelCase="<s>" , __UpperCAmelCase="</s>" , __UpperCAmelCase="<unk>" , __UpperCAmelCase="<sep>" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="<cls>" , __UpperCAmelCase="<mask>" , __UpperCAmelCase=["<eop>", "<eod>"] , **__UpperCAmelCase , ) ->List[str]: # Mask token behave like a normal word, i.e. include the space before it a_ = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase) if isinstance(__UpperCAmelCase , __UpperCAmelCase) else mask_token super().__init__( vocab_file=__UpperCAmelCase , tokenizer_file=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , additional_special_tokens=__UpperCAmelCase , **__UpperCAmelCase , ) a_ = 3 a_ = do_lower_case a_ = remove_space a_ = keep_accents a_ = vocab_file a_ = False if not self.vocab_file else True def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->List[int]: a_ = [self.sep_token_id] a_ = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->List[int]: a_ = [self.sep_token_id] a_ = [2] if token_ids_a is None: return len(token_ids_a + sep) * [0] + cls_segment_id return len(token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] + cls_segment_id def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None) ->Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(__UpperCAmelCase): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''') return a_ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(__UpperCAmelCase): copyfile(self.vocab_file , __UpperCAmelCase) return (out_vocab_file,)
364
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class snake_case ( SCREAMING_SNAKE_CASE_ ): a_ : Tuple = """audio-spectrogram-transformer""" def __init__( self , __UpperCAmelCase=7_68 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=30_72 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-12 , __UpperCAmelCase=16 , __UpperCAmelCase=True , __UpperCAmelCase=10 , __UpperCAmelCase=10 , __UpperCAmelCase=10_24 , __UpperCAmelCase=1_28 , **__UpperCAmelCase , ) ->str: super().__init__(**__UpperCAmelCase) a_ = hidden_size a_ = num_hidden_layers a_ = num_attention_heads a_ = intermediate_size a_ = hidden_act a_ = hidden_dropout_prob a_ = attention_probs_dropout_prob a_ = initializer_range a_ = layer_norm_eps a_ = patch_size a_ = qkv_bias a_ = frequency_stride a_ = time_stride a_ = max_length a_ = num_mel_bins
303
0
"""simple docstring""" import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin snake_case_ = get_tests_dir("""fixtures/test_sentencepiece_bpe.model""") class A_ ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCamelCase = BartphoTokenizer __UpperCamelCase = False __UpperCamelCase = True def UpperCAmelCase__ ( self :Tuple ) -> Any: super().setUp() UpperCAmelCase = ['▁This', '▁is', '▁a', '▁t', 'est'] UpperCAmelCase = dict(zip(a__ , range(len(a__ ) ) ) ) UpperCAmelCase = {'unk_token': '<unk>'} UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['monolingual_vocab_file'] ) with open(self.monolingual_vocab_file , 'w' , encoding='utf-8' ) as fp: for token in vocab_tokens: fp.write(f"""{token} {vocab_tokens[token]}\n""" ) UpperCAmelCase = BartphoTokenizer(a__ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase__ ( self :str , **lowercase_ :str ) -> str: kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **a__ ) def UpperCAmelCase__ ( self :str , lowercase_ :Any ) -> int: UpperCAmelCase = 'This is a là test' UpperCAmelCase = 'This is a<unk><unk> test' return input_text, output_text def UpperCAmelCase__ ( self :Dict ) -> List[Any]: UpperCAmelCase = BartphoTokenizer(a__ , self.monolingual_vocab_file , **self.special_tokens_map ) UpperCAmelCase = 'This is a là test' UpperCAmelCase = '▁This ▁is ▁a ▁l à ▁t est'.split() UpperCAmelCase = tokenizer.tokenize(a__ ) self.assertListEqual(a__ , a__ ) UpperCAmelCase = tokens + [tokenizer.unk_token] UpperCAmelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(a__ ) , a__ )
78
import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE__ : def __init__(self : Any , a__ : Union[str, Any] , a__ : int=13 , a__ : int=7 , a__ : Optional[Any]=True , a__ : Optional[int]=True , a__ : Any=True , a__ : str=True , a__ : List[Any]=99 , a__ : Any=24 , a__ : List[str]=2 , a__ : Optional[int]=6 , a__ : int=37 , a__ : List[str]="gelu" , a__ : List[Any]=0.1 , a__ : Optional[int]=0.1 , a__ : Union[str, Any]=512 , a__ : List[str]=16 , a__ : Optional[int]=2 , a__ : Union[str, Any]=0.0_2 , a__ : str=3 , a__ : Optional[Any]=None , a__ : Any=1000 , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = scope __snake_case = range_bbox def a (self : Optional[int] ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: __snake_case = bbox[i, j, 3] __snake_case = bbox[i, j, 1] __snake_case = t if bbox[i, j, 2] < bbox[i, j, 0]: __snake_case = bbox[i, j, 2] __snake_case = bbox[i, j, 0] __snake_case = t __snake_case = None if self.use_input_mask: __snake_case = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __snake_case = None if self.use_token_type_ids: __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def a (self : List[str] ): """simple docstring""" return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def a (self : List[Any] , a__ : List[Any] , a__ : Optional[Any] , a__ : List[str] , a__ : int , a__ : Optional[int] , a__ : str , a__ : Optional[int] , ): """simple docstring""" __snake_case = LiltModel(config=a__ ) model.to(a__ ) model.eval() __snake_case = model(a__ , bbox=a__ , attention_mask=a__ , token_type_ids=a__ ) __snake_case = model(a__ , bbox=a__ , token_type_ids=a__ ) __snake_case = model(a__ , bbox=a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a (self : Any , a__ : Tuple , a__ : Dict , a__ : Optional[int] , a__ : Dict , a__ : Union[str, Any] , a__ : str , a__ : Tuple , ): """simple docstring""" __snake_case = self.num_labels __snake_case = LiltForTokenClassification(config=a__ ) model.to(a__ ) model.eval() __snake_case = model( a__ , bbox=a__ , attention_mask=a__ , token_type_ids=a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a (self : int , a__ : Optional[Any] , a__ : int , a__ : int , a__ : Optional[Any] , a__ : Tuple , a__ : Union[str, Any] , a__ : str , ): """simple docstring""" __snake_case = LiltForQuestionAnswering(config=a__ ) model.to(a__ ) model.eval() __snake_case = model( a__ , bbox=a__ , attention_mask=a__ , token_type_ids=a__ , start_positions=a__ , end_positions=a__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a (self : Tuple ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = { '''input_ids''': input_ids, '''bbox''': bbox, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A_ : List[Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) A_ : Any = ( { 'feature-extraction': LiltModel, 'question-answering': LiltForQuestionAnswering, 'text-classification': LiltForSequenceClassification, 'token-classification': LiltForTokenClassification, 'zero-shot': LiltForSequenceClassification, } if is_torch_available() else {} ) A_ : Optional[int] = False A_ : List[Any] = False def a (self : Dict , a__ : Tuple , a__ : Tuple , a__ : Tuple , a__ : Union[str, Any] , a__ : Any ): """simple docstring""" return True def a (self : Union[str, Any] ): """simple docstring""" __snake_case = LiltModelTester(self ) __snake_case = ConfigTester(self , config_class=a__ , hidden_size=37 ) def a (self : Optional[int] ): """simple docstring""" self.config_tester.run_common_tests() def a (self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def a (self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __snake_case = type self.model_tester.create_and_check_model(*a__ ) def a (self : Optional[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a__ ) def a (self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a__ ) @slow def a (self : Optional[int] ): """simple docstring""" for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = LiltModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) @require_torch @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def a (self : Tuple ): """simple docstring""" __snake_case = LiltModel.from_pretrained('''SCUT-DLVCLab/lilt-roberta-en-base''' ).to(a__ ) __snake_case = torch.tensor([[1, 2]] , device=a__ ) __snake_case = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=a__ ) # forward pass with torch.no_grad(): __snake_case = model(input_ids=a__ , bbox=a__ ) __snake_case = torch.Size([1, 2, 768] ) __snake_case = torch.tensor( [[-0.0_6_5_3, 0.0_9_5_0, -0.0_0_6_1], [-0.0_5_4_5, 0.0_9_2_6, -0.0_3_2_4]] , device=a__ , ) self.assertTrue(outputs.last_hidden_state.shape , a__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , a__ , atol=1E-3 ) )
24
0
"""simple docstring""" _lowercase : Tuple = { 'A': '.-', 'B': '-...', 'C': '-.-.', 'D': '-..', 'E': '.', 'F': '..-.', 'G': '--.', 'H': '....', 'I': '..', 'J': '.---', 'K': '-.-', 'L': '.-..', 'M': '--', 'N': '-.', 'O': '---', 'P': '.--.', 'Q': '--.-', 'R': '.-.', 'S': '...', 'T': '-', 'U': '..-', 'V': '...-', 'W': '.--', 'X': '-..-', 'Y': '-.--', 'Z': '--..', '1': '.----', '2': '..---', '3': '...--', '4': '....-', '5': '.....', '6': '-....', '7': '--...', '8': '---..', '9': '----.', '0': '-----', '&': '.-...', '@': '.--.-.', ':': '---...', ',': '--..--', '.': '.-.-.-', '\'': '.----.', '"': '.-..-.', '?': '..--..', '/': '-..-.', '=': '-...-', '+': '.-.-.', '-': '-....-', '(': '-.--.', ')': '-.--.-', '!': '-.-.--', ' ': '/' } # Exclamation mark is not in ITU-R recommendation # fmt: on _lowercase : Dict = {value: key for key, value in MORSE_CODE_DICT.items()} def lowercase__ ( snake_case_ :str ): return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowercase__ ( snake_case_ :str ): return "".join(REVERSE_DICT[char] for char in message.split() ) def lowercase__ ( ): __UpperCAmelCase = '''Morse code here!''' print(snake_case_ ) __UpperCAmelCase = encrypt(snake_case_ ) print(snake_case_ ) __UpperCAmelCase = decrypt(snake_case_ ) print(snake_case_ ) if __name__ == "__main__": main()
355
"""simple docstring""" import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class _UpperCAmelCase ( _lowerCAmelCase , unittest.TestCase ): a__ : Tuple = RoCBertTokenizer a__ : List[Any] = None a__ : List[Any] = False a__ : Dict = True a__ : int = filter_non_english def a ( self : Optional[int] ): super().setUp() __UpperCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''你''', '''好''', '''是''', '''谁''', '''a''', '''b''', '''c''', '''d'''] __UpperCAmelCase = {} __UpperCAmelCase = {} for i, value in enumerate(_lowercase ): __UpperCAmelCase = i __UpperCAmelCase = i __UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''word_shape_file'''] ) __UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''word_pronunciation_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) with open(self.word_shape_file , '''w''' , encoding='''utf-8''' ) as word_shape_writer: json.dump(_lowercase , _lowercase , ensure_ascii=_lowercase ) with open(self.word_pronunciation_file , '''w''' , encoding='''utf-8''' ) as word_pronunciation_writer: json.dump(_lowercase , _lowercase , ensure_ascii=_lowercase ) def a ( self : Optional[Any] ): __UpperCAmelCase = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) __UpperCAmelCase = tokenizer.tokenize('''你好[SEP]你是谁''' ) self.assertListEqual(_lowercase , ['''你''', '''好''', '''[SEP]''', '''你''', '''是''', '''谁'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowercase ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(_lowercase ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(_lowercase ) , [5, 6, 2, 5, 7, 8] ) def a ( self : List[Any] ): __UpperCAmelCase = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''' ) , ['''ah''', '''\u535A''', '''\u63A8''', '''zz'''] ) def a ( self : Union[str, Any] ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''hello''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def a ( self : Optional[Any] ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''h\u00E9llo'''] ) def a ( self : List[Any] ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def a ( self : Optional[Any] ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''hello'''] ) def a ( self : Optional[int] ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def a ( self : Union[str, Any] ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def a ( self : Any ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase , strip_accents=_lowercase ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) , ['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def a ( self : int ): __UpperCAmelCase = RoCBertBasicTokenizer(do_lower_case=_lowercase , never_split=['''[UNK]'''] ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''' ) , ['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]'''] ) def a ( self : Optional[Any] ): __UpperCAmelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] __UpperCAmelCase = {} for i, token in enumerate(_lowercase ): __UpperCAmelCase = i __UpperCAmelCase = RoCBertWordpieceTokenizer(vocab=_lowercase , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''unwanted running''' ) , ['''un''', '''##want''', '''##ed''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.tokenize('''unwantedX running''' ) , ['''[UNK]''', '''runn''', '''##ing'''] ) def a ( self : Union[str, Any] ): self.assertTrue(_is_whitespace(''' ''' ) ) self.assertTrue(_is_whitespace('''\t''' ) ) self.assertTrue(_is_whitespace('''\r''' ) ) self.assertTrue(_is_whitespace('''\n''' ) ) self.assertTrue(_is_whitespace('''\u00A0''' ) ) self.assertFalse(_is_whitespace('''A''' ) ) self.assertFalse(_is_whitespace('''-''' ) ) def a ( self : Dict ): self.assertTrue(_is_control('''\u0005''' ) ) self.assertFalse(_is_control('''A''' ) ) self.assertFalse(_is_control(''' ''' ) ) self.assertFalse(_is_control('''\t''' ) ) self.assertFalse(_is_control('''\r''' ) ) def a ( self : Optional[int] ): self.assertTrue(_is_punctuation('''-''' ) ) self.assertTrue(_is_punctuation('''$''' ) ) self.assertTrue(_is_punctuation('''`''' ) ) self.assertTrue(_is_punctuation('''.''' ) ) self.assertFalse(_is_punctuation('''A''' ) ) self.assertFalse(_is_punctuation(''' ''' ) ) def a ( self : Tuple ): __UpperCAmelCase = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_lowercase ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) if self.test_rust_tokenizer: __UpperCAmelCase = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(_lowercase ) for t in ['''Test''', '''\xad''', '''test''']] , [['''[UNK]'''], [], ['''[UNK]''']] ) def a ( self : Optional[int] ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = F'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' __UpperCAmelCase = tokenizer_r.encode_plus( _lowercase , return_attention_mask=_lowercase , return_token_type_ids=_lowercase , return_offsets_mapping=_lowercase , add_special_tokens=_lowercase , ) __UpperCAmelCase = tokenizer_r.do_lower_case if hasattr(_lowercase , '''do_lower_case''' ) else False __UpperCAmelCase = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''A'''), ((1, 2), ''','''), ((3, 5), '''na'''), ((5, 6), '''##ï'''), ((6, 8), '''##ve'''), ((9, 15), tokenizer_r.mask_token), ((16, 21), '''Allen'''), ((21, 23), '''##NL'''), ((23, 24), '''##P'''), ((25, 33), '''sentence'''), ((33, 34), '''.'''), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''a'''), ((1, 2), ''','''), ((3, 8), '''naive'''), ((9, 15), tokenizer_r.mask_token), ((16, 21), '''allen'''), ((21, 23), '''##nl'''), ((23, 24), '''##p'''), ((25, 33), '''sentence'''), ((33, 34), '''.'''), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['''input_ids'''] ) ) self.assertEqual([e[0] for e in expected_results] , tokens['''offset_mapping'''] ) def a ( self : Dict ): __UpperCAmelCase = ['''的''', '''人''', '''有'''] __UpperCAmelCase = ''''''.join(_lowercase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCAmelCase = True __UpperCAmelCase = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = tokenizer_p.encode(_lowercase , add_special_tokens=_lowercase ) __UpperCAmelCase = tokenizer_r.encode(_lowercase , add_special_tokens=_lowercase ) __UpperCAmelCase = tokenizer_r.convert_ids_to_tokens(_lowercase ) __UpperCAmelCase = tokenizer_p.convert_ids_to_tokens(_lowercase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_lowercase , _lowercase ) self.assertListEqual(_lowercase , _lowercase ) __UpperCAmelCase = False __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = tokenizer_r.encode(_lowercase , add_special_tokens=_lowercase ) __UpperCAmelCase = tokenizer_p.encode(_lowercase , add_special_tokens=_lowercase ) __UpperCAmelCase = tokenizer_r.convert_ids_to_tokens(_lowercase ) __UpperCAmelCase = tokenizer_p.convert_ids_to_tokens(_lowercase ) # it is expected that only the first Chinese character is not preceded by "##". __UpperCAmelCase = [ F'''##{token}''' if idx != 0 else token for idx, token in enumerate(_lowercase ) ] self.assertListEqual(_lowercase , _lowercase ) self.assertListEqual(_lowercase , _lowercase ) @slow def a ( self : List[Any] ): __UpperCAmelCase = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) __UpperCAmelCase = tokenizer.encode('''你好''' , add_special_tokens=_lowercase ) __UpperCAmelCase = tokenizer.encode('''你是谁''' , add_special_tokens=_lowercase ) __UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(_lowercase ) __UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(_lowercase , _lowercase ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def a ( self : List[str] ): __UpperCAmelCase = self.get_tokenizers(do_lower_case=_lowercase ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): __UpperCAmelCase = '''你好,你是谁''' __UpperCAmelCase = tokenizer.tokenize(_lowercase ) __UpperCAmelCase = tokenizer.convert_tokens_to_ids(_lowercase ) __UpperCAmelCase = tokenizer.convert_tokens_to_shape_ids(_lowercase ) __UpperCAmelCase = tokenizer.convert_tokens_to_pronunciation_ids(_lowercase ) __UpperCAmelCase = tokenizer.prepare_for_model( _lowercase , _lowercase , _lowercase , add_special_tokens=_lowercase ) __UpperCAmelCase = tokenizer.encode_plus(_lowercase , add_special_tokens=_lowercase ) self.assertEqual(_lowercase , _lowercase )
86
0
'''simple docstring''' import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class a__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" __lowerCamelCase = logging.get_logger() # the current default level is logging.WARNING __lowerCamelCase = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(a ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" __lowerCamelCase = logging.get_verbosity() __lowerCamelCase = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) __lowerCamelCase = '''Testing 1, 2, 3''' # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(a ) as cl: logger.warning(a ) self.assertEqual(cl.out , msg + '''\n''' ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(a ) as cl: logger.warning(a ) self.assertEqual(cl.out , '''''' ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(a ) as cl: logger.warning(a ) self.assertEqual(cl.out , msg + '''\n''' ) # restore to the original level logging.set_verbosity(a ) @mockenv(TRANSFORMERS_VERBOSITY='''error''' ) def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" transformers.utils.logging._reset_library_root_logger() # this action activates the env var __lowerCamelCase = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) __lowerCamelCase = os.getenv('''TRANSFORMERS_VERBOSITY''' , a ) __lowerCamelCase = logging.log_levels[env_level_str] __lowerCamelCase = logging.get_verbosity() self.assertEqual( a , a , f"""TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}""" , ) # restore to the original level __lowerCamelCase = '''''' transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY='''super-error''' ) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" transformers.utils.logging._reset_library_root_logger() __lowerCamelCase = logging.logging.getLogger() with CaptureLogger(a ) as cl: # this action activates the env var logging.get_logger('''transformers.models.bart.tokenization_bart''' ) self.assertIn('''Unknown option TRANSFORMERS_VERBOSITY=super-error''' , cl.out ) # no need to restore as nothing was changed def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" transformers.utils.logging._reset_library_root_logger() __lowerCamelCase = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) __lowerCamelCase = '''Testing 1, 2, 3''' with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''1''' ): # nothing should be logged as env var disables this method with CaptureLogger(a ) as cl: logger.warning_advice(a ) self.assertEqual(cl.out , '''''' ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''''' ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(a ) as cl: logger.warning_advice(a ) self.assertEqual(cl.out , msg + '''\n''' ) def __lowerCAmelCase ( ) -> List[str]: disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
67
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a_ : List[Any] = get_logger(__name__) class _snake_case ( enum.Enum ): _lowercase : Any = '''all_checks''' _lowercase : str = '''basic_checks''' _lowercase : str = '''no_checks''' class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None): if expected_checksums is None: logger.info('Unable to verify checksums.') return if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise ExpectedMoreDownloadedFiles(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise UnexpectedDownloadedFile(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) SCREAMING_SNAKE_CASE = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] SCREAMING_SNAKE_CASE = ' for ' + verification_name if verification_name is not None else '' if len(_UpperCAmelCase) > 0: raise NonMatchingChecksumError( F'''Checksums didn\'t match{for_verification_name}:\n''' F'''{bad_urls}\n''' 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error') logger.info('All the checksums matched successfully' + for_verification_name) class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase): if expected_splits is None: logger.info('Unable to verify splits sizes.') return if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise ExpectedMoreSplits(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise UnexpectedSplits(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) SCREAMING_SNAKE_CASE = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_UpperCAmelCase) > 0: raise NonMatchingSplitsSizesError(str(_UpperCAmelCase)) logger.info('All the splits matched successfully.') def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase = True): if record_checksum: SCREAMING_SNAKE_CASE = shaaaa() with open(_UpperCAmelCase , 'rb') as f: for chunk in iter(lambda: f.read(1 << 20) , B''): m.update(_UpperCAmelCase) SCREAMING_SNAKE_CASE = m.hexdigest() else: SCREAMING_SNAKE_CASE = None return {"num_bytes": os.path.getsize(_UpperCAmelCase), "checksum": checksum} def lowerCamelCase__ (_UpperCAmelCase): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
137
0
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow lowercase_ = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class SCREAMING_SNAKE_CASE (unittest.TestCase ): def SCREAMING_SNAKE_CASE_ ( self : str , a : Path , a : Union[str, None] = None , a : Union[List[str], None] = None , a : Union[str, List[str], None] = None , a : bool = True , )-> Tuple: """simple docstring""" lowercase__ = [file for file in os.listdir(a ) if os.path.isfile(os.path.join(a , a ) )] if identifier is not None: lowercase__ = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(a , a ): for n_ in n_identifier: lowercase__ = [file for file in files if n_ not in file] else: lowercase__ = [file for file in files if n_identifier not in file] lowercase__ = ignore_files or [] ignore_files.append('__init__.py' ) lowercase__ = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' , a ) if only_modules: lowercase__ = file.split('.' )[0] try: lowercase__ = getattr(a , a ) lowercase__ = doctest.DocTestSuite(a ) lowercase__ = unittest.TextTestRunner().run(a ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(f"""{module_identifier} is not a module.""" ) else: lowercase__ = doctest.testfile(str('..' / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def SCREAMING_SNAKE_CASE_ ( self : Dict )-> Tuple: """simple docstring""" lowercase__ = Path('src/transformers' ) lowercase__ = 'modeling' lowercase__ = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(a , identifier=a , ignore_files=a ) def SCREAMING_SNAKE_CASE_ ( self : List[str] )-> List[Any]: """simple docstring""" lowercase__ = Path('src/transformers' ) lowercase__ = 'tokenization' self.analyze_directory(a , identifier=a ) def SCREAMING_SNAKE_CASE_ ( self : str )-> int: """simple docstring""" lowercase__ = Path('src/transformers' ) lowercase__ = 'configuration' self.analyze_directory(a , identifier=a ) def SCREAMING_SNAKE_CASE_ ( self : List[str] )-> Optional[Any]: """simple docstring""" lowercase__ = Path('src/transformers' ) lowercase__ = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(a , n_identifier=a ) def SCREAMING_SNAKE_CASE_ ( self : Any )-> Dict: """simple docstring""" lowercase__ = Path('docs/source' ) lowercase__ = ['favicon.ico'] self.analyze_directory(a , ignore_files=a , only_modules=a )
360
from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def __UpperCamelCase () -> str: import os as original_os from os import path as original_path from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join lowercase__ = '__test_patch_submodule_mock__' with patch_submodule(_test_patching , 'os.path.join' , _SCREAMING_SNAKE_CASE ): # Every way to access os.path.join must be patched, and the rest must stay untouched # check os.path.join assert isinstance(_test_patching.os , _PatchedModuleObj ) assert isinstance(_test_patching.os.path , _PatchedModuleObj ) assert _test_patching.os.path.join is mock # check path.join assert isinstance(_test_patching.path , _PatchedModuleObj ) assert _test_patching.path.join is mock # check join assert _test_patching.join is mock # check that the other attributes are untouched assert _test_patching.os.rename is original_rename assert _test_patching.path.dirname is original_dirname assert _test_patching.os.path.dirname is original_dirname # Even renamed modules or objects must be patched # check renamed_os.path.join assert isinstance(_test_patching.renamed_os , _PatchedModuleObj ) assert isinstance(_test_patching.renamed_os.path , _PatchedModuleObj ) assert _test_patching.renamed_os.path.join is mock # check renamed_path.join assert isinstance(_test_patching.renamed_path , _PatchedModuleObj ) assert _test_patching.renamed_path.join is mock # check renamed_join assert _test_patching.renamed_join is mock # check that the other attributes are untouched assert _test_patching.renamed_os.rename is original_rename assert _test_patching.renamed_path.dirname is original_dirname assert _test_patching.renamed_os.path.dirname is original_dirname # check that everthing is back to normal when the patch is over assert _test_patching.os is original_os assert _test_patching.path is original_path assert _test_patching.join is original_join assert _test_patching.renamed_os is original_os assert _test_patching.renamed_path is original_path assert _test_patching.renamed_join is original_join def __UpperCamelCase () -> Any: assert _test_patching.open is open lowercase__ = '__test_patch_submodule_builtin_mock__' # _test_patching has "open" in its globals assert _test_patching.open is open with patch_submodule(_test_patching , 'open' , _SCREAMING_SNAKE_CASE ): assert _test_patching.open is mock # check that everthing is back to normal when the patch is over assert _test_patching.open is open def __UpperCamelCase () -> List[str]: # pandas.read_csv is not present in _test_patching lowercase__ = '__test_patch_submodule_missing_mock__' with patch_submodule(_test_patching , 'pandas.read_csv' , _SCREAMING_SNAKE_CASE ): pass def __UpperCamelCase () -> List[str]: # builtin should always be mocked even if they're not in the globals # in case they're loaded at one point lowercase__ = '__test_patch_submodule_missing_builtin_mock__' # _test_patching doesn't have "len" in its globals assert getattr(_test_patching , 'len' , _SCREAMING_SNAKE_CASE ) is None with patch_submodule(_test_patching , 'len' , _SCREAMING_SNAKE_CASE ): assert _test_patching.len is mock assert _test_patching.len is len def __UpperCamelCase () -> List[str]: lowercase__ = '__test_patch_submodule_start_and_stop_mock__' lowercase__ = patch_submodule(_test_patching , 'open' , _SCREAMING_SNAKE_CASE ) assert _test_patching.open is open patch.start() assert _test_patching.open is mock patch.stop() assert _test_patching.open is open def __UpperCamelCase () -> Optional[int]: from os import rename as original_rename from os.path import dirname as original_dirname from os.path import join as original_join lowercase__ = '__test_patch_submodule_successive_join__' lowercase__ = '__test_patch_submodule_successive_dirname__' lowercase__ = '__test_patch_submodule_successive_rename__' assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename with patch_submodule(_test_patching , 'os.path.join' , _SCREAMING_SNAKE_CASE ): with patch_submodule(_test_patching , 'os.rename' , _SCREAMING_SNAKE_CASE ): with patch_submodule(_test_patching , 'os.path.dirname' , _SCREAMING_SNAKE_CASE ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename # try another order with patch_submodule(_test_patching , 'os.rename' , _SCREAMING_SNAKE_CASE ): with patch_submodule(_test_patching , 'os.path.join' , _SCREAMING_SNAKE_CASE ): with patch_submodule(_test_patching , 'os.path.dirname' , _SCREAMING_SNAKE_CASE ): assert _test_patching.os.path.join is mock_join assert _test_patching.os.path.dirname is mock_dirname assert _test_patching.os.rename is mock_rename assert _test_patching.os.path.join is original_join assert _test_patching.os.path.dirname is original_dirname assert _test_patching.os.rename is original_rename def __UpperCamelCase () -> Optional[Any]: lowercase__ = '__test_patch_submodule_doesnt_exist_mock__' with patch_submodule(_test_patching , '__module_that_doesn_exist__.__attribute_that_doesn_exist__' , _SCREAMING_SNAKE_CASE ): pass with patch_submodule(_test_patching , 'os.__attribute_that_doesn_exist__' , _SCREAMING_SNAKE_CASE ): pass
269
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : int = { 'microsoft/biogpt': 'https://huggingface.co/microsoft/biogpt/resolve/main/config.json', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = """biogpt""" def __init__( self : List[str] , UpperCamelCase__ : Optional[Any]=4_2384 , UpperCamelCase__ : Union[str, Any]=1024 , UpperCamelCase__ : Any=24 , UpperCamelCase__ : Union[str, Any]=16 , UpperCamelCase__ : Tuple=4096 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : Dict=0.1 , UpperCamelCase__ : Dict=0.1 , UpperCamelCase__ : str=1024 , UpperCamelCase__ : Any=0.02 , UpperCamelCase__ : List[str]=1E-12 , UpperCamelCase__ : Dict=True , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : Union[str, Any]=0.0 , UpperCamelCase__ : Any=0.0 , UpperCamelCase__ : List[str]=1 , UpperCamelCase__ : Dict=0 , UpperCamelCase__ : List[str]=2 , **UpperCamelCase__ : Optional[int] , ) -> Tuple: """simple docstring""" __magic_name__ = vocab_size __magic_name__ = max_position_embeddings __magic_name__ = hidden_size __magic_name__ = num_hidden_layers __magic_name__ = num_attention_heads __magic_name__ = intermediate_size __magic_name__ = hidden_act __magic_name__ = hidden_dropout_prob __magic_name__ = attention_probs_dropout_prob __magic_name__ = initializer_range __magic_name__ = layer_norm_eps __magic_name__ = scale_embedding __magic_name__ = use_cache __magic_name__ = layerdrop __magic_name__ = activation_dropout super().__init__(pad_token_id=UpperCamelCase__ , bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ )
88
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = 42 class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : List[Any]=3 , UpperCamelCase__ : str=3 , UpperCamelCase__ : List[Any]=("DownEncoderBlock2D",) , UpperCamelCase__ : Optional[Any]=(64,) , UpperCamelCase__ : Optional[Any]=2 , UpperCamelCase__ : Union[str, Any]=32 , UpperCamelCase__ : Optional[Any]="silu" , UpperCamelCase__ : List[str]=True , ) -> str: """simple docstring""" super().__init__() __magic_name__ = layers_per_block __magic_name__ = torch.nn.Convad( UpperCamelCase__ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) __magic_name__ = None __magic_name__ = nn.ModuleList([] ) # down __magic_name__ = block_out_channels[0] for i, down_block_type in enumerate(UpperCamelCase__ ): __magic_name__ = output_channel __magic_name__ = block_out_channels[i] __magic_name__ = i == len(UpperCamelCase__ ) - 1 __magic_name__ = get_down_block( UpperCamelCase__ , num_layers=self.layers_per_block , in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=UpperCamelCase__ , resnet_groups=UpperCamelCase__ , attention_head_dim=UpperCamelCase__ , temb_channels=UpperCamelCase__ , ) self.down_blocks.append(UpperCamelCase__ ) # mid __magic_name__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=UpperCamelCase__ , output_scale_factor=1 , resnet_time_scale_shift="""default""" , attention_head_dim=block_out_channels[-1] , resnet_groups=UpperCamelCase__ , temb_channels=UpperCamelCase__ , ) # out __magic_name__ = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=UpperCamelCase__ , eps=1E-6 ) __magic_name__ = nn.SiLU() __magic_name__ = 2 * out_channels if double_z else out_channels __magic_name__ = nn.Convad(block_out_channels[-1] , UpperCamelCase__ , 3 , padding=1 ) __magic_name__ = False def _lowercase ( self : List[str] , UpperCamelCase__ : Optional[Any] ) -> int: """simple docstring""" __magic_name__ = x __magic_name__ = self.conv_in(UpperCamelCase__ ) if self.training and self.gradient_checkpointing: def create_custom_forward(UpperCamelCase__ : int ): def custom_forward(*UpperCamelCase__ : str ): return module(*UpperCamelCase__ ) return custom_forward # down if is_torch_version(""">=""" , """1.11.0""" ): for down_block in self.down_blocks: __magic_name__ = torch.utils.checkpoint.checkpoint( create_custom_forward(UpperCamelCase__ ) , UpperCamelCase__ , use_reentrant=UpperCamelCase__ ) # middle __magic_name__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , UpperCamelCase__ , use_reentrant=UpperCamelCase__ ) else: for down_block in self.down_blocks: __magic_name__ = torch.utils.checkpoint.checkpoint(create_custom_forward(UpperCamelCase__ ) , UpperCamelCase__ ) # middle __magic_name__ = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , UpperCamelCase__ ) else: # down for down_block in self.down_blocks: __magic_name__ = down_block(UpperCamelCase__ ) # middle __magic_name__ = self.mid_block(UpperCamelCase__ ) # post-process __magic_name__ = self.conv_norm_out(UpperCamelCase__ ) __magic_name__ = self.conv_act(UpperCamelCase__ ) __magic_name__ = self.conv_out(UpperCamelCase__ ) return sample class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : int=3 , UpperCamelCase__ : Dict=3 , UpperCamelCase__ : List[Any]=("UpDecoderBlock2D",) , UpperCamelCase__ : List[Any]=(64,) , UpperCamelCase__ : Dict=2 , UpperCamelCase__ : int=32 , UpperCamelCase__ : Optional[int]="silu" , UpperCamelCase__ : Tuple="group" , ) -> Dict: """simple docstring""" super().__init__() __magic_name__ = layers_per_block __magic_name__ = nn.Convad( UpperCamelCase__ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) __magic_name__ = None __magic_name__ = nn.ModuleList([] ) __magic_name__ = in_channels if norm_type == """spatial""" else None # mid __magic_name__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=UpperCamelCase__ , output_scale_factor=1 , resnet_time_scale_shift="""default""" if norm_type == """group""" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=UpperCamelCase__ , temb_channels=UpperCamelCase__ , ) # up __magic_name__ = list(reversed(UpperCamelCase__ ) ) __magic_name__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(UpperCamelCase__ ): __magic_name__ = output_channel __magic_name__ = reversed_block_out_channels[i] __magic_name__ = i == len(UpperCamelCase__ ) - 1 __magic_name__ = get_up_block( UpperCamelCase__ , num_layers=self.layers_per_block + 1 , in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , prev_output_channel=UpperCamelCase__ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=UpperCamelCase__ , resnet_groups=UpperCamelCase__ , attention_head_dim=UpperCamelCase__ , temb_channels=UpperCamelCase__ , resnet_time_scale_shift=UpperCamelCase__ , ) self.up_blocks.append(UpperCamelCase__ ) __magic_name__ = output_channel # out if norm_type == "spatial": __magic_name__ = SpatialNorm(block_out_channels[0] , UpperCamelCase__ ) else: __magic_name__ = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=UpperCamelCase__ , eps=1E-6 ) __magic_name__ = nn.SiLU() __magic_name__ = nn.Convad(block_out_channels[0] , UpperCamelCase__ , 3 , padding=1 ) __magic_name__ = False def _lowercase ( self : str , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Tuple=None ) -> Tuple: """simple docstring""" __magic_name__ = z __magic_name__ = self.conv_in(UpperCamelCase__ ) __magic_name__ = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(UpperCamelCase__ : Optional[int] ): def custom_forward(*UpperCamelCase__ : int ): return module(*UpperCamelCase__ ) return custom_forward if is_torch_version(""">=""" , """1.11.0""" ): # middle __magic_name__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , UpperCamelCase__ , UpperCamelCase__ , use_reentrant=UpperCamelCase__ ) __magic_name__ = sample.to(UpperCamelCase__ ) # up for up_block in self.up_blocks: __magic_name__ = torch.utils.checkpoint.checkpoint( create_custom_forward(UpperCamelCase__ ) , UpperCamelCase__ , UpperCamelCase__ , use_reentrant=UpperCamelCase__ ) else: # middle __magic_name__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = sample.to(UpperCamelCase__ ) # up for up_block in self.up_blocks: __magic_name__ = torch.utils.checkpoint.checkpoint(create_custom_forward(UpperCamelCase__ ) , UpperCamelCase__ , UpperCamelCase__ ) else: # middle __magic_name__ = self.mid_block(UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = sample.to(UpperCamelCase__ ) # up for up_block in self.up_blocks: __magic_name__ = up_block(UpperCamelCase__ , UpperCamelCase__ ) # post-process if latent_embeds is None: __magic_name__ = self.conv_norm_out(UpperCamelCase__ ) else: __magic_name__ = self.conv_norm_out(UpperCamelCase__ , UpperCamelCase__ ) __magic_name__ = self.conv_act(UpperCamelCase__ ) __magic_name__ = self.conv_out(UpperCamelCase__ ) return sample class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : str , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Dict="random" , UpperCamelCase__ : List[Any]=False , UpperCamelCase__ : Dict=True ) -> Optional[Any]: """simple docstring""" super().__init__() __magic_name__ = n_e __magic_name__ = vq_embed_dim __magic_name__ = beta __magic_name__ = legacy __magic_name__ = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) __magic_name__ = remap if self.remap is not None: self.register_buffer("""used""" , torch.tensor(np.load(self.remap ) ) ) __magic_name__ = self.used.shape[0] __magic_name__ = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": __magic_name__ = self.re_embed __magic_name__ = self.re_embed + 1 print( F'''Remapping {self.n_e} indices to {self.re_embed} indices. ''' F'''Using {self.unknown_index} for unknown indices.''' ) else: __magic_name__ = n_e __magic_name__ = sane_index_shape def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : Dict ) -> Union[str, Any]: """simple docstring""" __magic_name__ = inds.shape assert len(UpperCamelCase__ ) > 1 __magic_name__ = inds.reshape(ishape[0] , -1 ) __magic_name__ = self.used.to(UpperCamelCase__ ) __magic_name__ = (inds[:, :, None] == used[None, None, ...]).long() __magic_name__ = match.argmax(-1 ) __magic_name__ = match.sum(2 ) < 1 if self.unknown_index == "random": __magic_name__ = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: __magic_name__ = self.unknown_index return new.reshape(UpperCamelCase__ ) def _lowercase ( self : Optional[Any] , UpperCamelCase__ : str ) -> Tuple: """simple docstring""" __magic_name__ = inds.shape assert len(UpperCamelCase__ ) > 1 __magic_name__ = inds.reshape(ishape[0] , -1 ) __magic_name__ = self.used.to(UpperCamelCase__ ) if self.re_embed > self.used.shape[0]: # extra token __magic_name__ = 0 # simply set to zero __magic_name__ = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , UpperCamelCase__ ) return back.reshape(UpperCamelCase__ ) def _lowercase ( self : List[str] , UpperCamelCase__ : List[str] ) -> List[str]: """simple docstring""" __magic_name__ = z.permute(0 , 2 , 3 , 1 ).contiguous() __magic_name__ = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z __magic_name__ = torch.argmin(torch.cdist(UpperCamelCase__ , self.embedding.weight ) , dim=1 ) __magic_name__ = self.embedding(UpperCamelCase__ ).view(z.shape ) __magic_name__ = None __magic_name__ = None # compute loss for embedding if not self.legacy: __magic_name__ = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: __magic_name__ = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients __magic_name__ = z + (z_q - z).detach() # reshape back to match original input shape __magic_name__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: __magic_name__ = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis __magic_name__ = self.remap_to_used(UpperCamelCase__ ) __magic_name__ = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: __magic_name__ = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Union[str, Any] ) -> int: """simple docstring""" if self.remap is not None: __magic_name__ = indices.reshape(shape[0] , -1 ) # add batch axis __magic_name__ = self.unmap_to_all(UpperCamelCase__ ) __magic_name__ = indices.reshape(-1 ) # flatten again # get quantized latent vectors __magic_name__ = self.embedding(UpperCamelCase__ ) if shape is not None: __magic_name__ = z_q.view(UpperCamelCase__ ) # reshape back to match original input shape __magic_name__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class UpperCAmelCase_ ( _A ): '''simple docstring''' def __init__( self : Dict , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Tuple=False ) -> Optional[int]: """simple docstring""" __magic_name__ = parameters __magic_name__ , __magic_name__ = torch.chunk(UpperCamelCase__ , 2 , dim=1 ) __magic_name__ = torch.clamp(self.logvar , -30.0 , 20.0 ) __magic_name__ = deterministic __magic_name__ = torch.exp(0.5 * self.logvar ) __magic_name__ = torch.exp(self.logvar ) if self.deterministic: __magic_name__ = __magic_name__ = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def _lowercase ( self : Tuple , UpperCamelCase__ : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" __magic_name__ = randn_tensor( self.mean.shape , generator=UpperCamelCase__ , device=self.parameters.device , dtype=self.parameters.dtype ) __magic_name__ = self.mean + self.std * sample return x def _lowercase ( self : Dict , UpperCamelCase__ : Optional[int]=None ) -> Any: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def _lowercase ( self : Optional[int] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Dict=[1, 2, 3] ) -> Optional[int]: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) __magic_name__ = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=UpperCamelCase__ ) def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" return self.mean
88
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule _SCREAMING_SNAKE_CASE = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
355
import re import string import numpy as np import datasets _SCREAMING_SNAKE_CASE = """ Returns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list. """ _SCREAMING_SNAKE_CASE = """ Args: predictions: List of predicted texts. references: List of reference texts. regexes_to_ignore: List, defaults to None. Regex expressions of characters to ignore when calculating the exact matches. Note: these regexes are removed from the input data before the changes based on the options below (e.g. ignore_case, ignore_punctuation, ignore_numbers) are applied. ignore_case: Boolean, defaults to False. If true, turns everything to lowercase so that capitalization differences are ignored. ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. Returns: exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive. Examples: >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results[\"exact_match\"], 1)) 25.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\"], ignore_case=True, ignore_punctuation=True) >>> print(round(results[\"exact_match\"], 1)) 50.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True) >>> print(round(results[\"exact_match\"], 1)) 75.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True) >>> print(round(results[\"exact_match\"], 1)) 100.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"The cat sat on the mat.\", \"Theaters are great.\", \"It's like comparing oranges and apples.\"] >>> preds = [\"The cat sat on the mat?\", \"Theaters are great.\", \"It's like comparing apples and oranges.\"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results[\"exact_match\"], 1)) 33.3 """ _SCREAMING_SNAKE_CASE = """ """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE_ ( datasets.Metric ): def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , reference_urls=[] , ) def UpperCAmelCase_ ( self : List[Any] , _A : Optional[Any] , _A : Optional[int] , _A : Optional[int]=None , _A : Dict=False , _A : Dict=False , _A : Optional[Any]=False , ) -> List[str]: """simple docstring""" if regexes_to_ignore is not None: for s in regexes_to_ignore: snake_case_ : List[str] = np.array([re.sub(_A , '' , _A ) for x in predictions] ) snake_case_ : int = np.array([re.sub(_A , '' , _A ) for x in references] ) else: snake_case_ : Optional[Any] = np.asarray(_A ) snake_case_ : Optional[Any] = np.asarray(_A ) if ignore_case: snake_case_ : int = np.char.lower(_A ) snake_case_ : List[str] = np.char.lower(_A ) if ignore_punctuation: snake_case_ : str = string.punctuation.maketrans('' , '' , string.punctuation ) snake_case_ : str = np.char.translate(_A , table=_A ) snake_case_ : Any = np.char.translate(_A , table=_A ) if ignore_numbers: snake_case_ : int = string.digits.maketrans('' , '' , string.digits ) snake_case_ : Tuple = np.char.translate(_A , table=_A ) snake_case_ : Optional[Any] = np.char.translate(_A , table=_A ) snake_case_ : Optional[Any] = predictions == references return {"exact_match": np.mean(_A ) * 100}
88
0
"""simple docstring""" from ....utils import logging _A = logging.get_logger(__name__) class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=2048 ): """simple docstring""" UpperCAmelCase__ : Tuple = config.__dict__ UpperCAmelCase__ : Union[str, Any] = modal_hidden_size if num_labels: UpperCAmelCase__ : Tuple = num_labels
171
def a__ ( snake_case = 1_000_000 ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = 1 __SCREAMING_SNAKE_CASE : Optional[Any] = 1 __SCREAMING_SNAKE_CASE : Optional[int] = {1: 1} for inputa in range(2 , snake_case ): __SCREAMING_SNAKE_CASE : Tuple = 0 __SCREAMING_SNAKE_CASE : Optional[Any] = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __SCREAMING_SNAKE_CASE : List[Any] = (3 * number) + 1 counter += 1 if inputa not in counters: __SCREAMING_SNAKE_CASE : str = counter if counter > pre_counter: __SCREAMING_SNAKE_CASE : Optional[int] = inputa __SCREAMING_SNAKE_CASE : str = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
303
0
"""simple docstring""" from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __A : int = logging.get_logger(__name__) class __UpperCamelCase ( __SCREAMING_SNAKE_CASE ): SCREAMING_SNAKE_CASE = ["pixel_values"] def __init__(self : int , __SCREAMING_SNAKE_CASE : Union[str, Any] = True , __SCREAMING_SNAKE_CASE : List[Any] = None , __SCREAMING_SNAKE_CASE : Dict = PILImageResampling.BICUBIC , __SCREAMING_SNAKE_CASE : Union[str, Any] = True , __SCREAMING_SNAKE_CASE : Optional[Any] = None , __SCREAMING_SNAKE_CASE : Optional[Any] = True , __SCREAMING_SNAKE_CASE : Tuple = 1 / 2_5_5 , __SCREAMING_SNAKE_CASE : List[Any] = True , __SCREAMING_SNAKE_CASE : Dict = IMAGENET_DEFAULT_MEAN , __SCREAMING_SNAKE_CASE : Optional[Any] = IMAGENET_DEFAULT_STD , **__SCREAMING_SNAKE_CASE : Optional[Any] , ): super().__init__(**_snake_case) A = size if size is not None else {"shortest_edge": 2_2_4} A = get_size_dict(_snake_case , default_to_square=_snake_case) A = crop_size if crop_size is not None else {"height": 2_2_4, "width": 2_2_4} A = get_size_dict(_snake_case , param_name="crop_size") A = do_resize A = size A = resample A = do_center_crop A = crop_size A = do_rescale A = rescale_factor A = do_normalize A = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN A = image_std if image_std is not None else IMAGENET_DEFAULT_STD def SCREAMING_SNAKE_CASE__ (self : List[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : List[str] = PILImageResampling.BICUBIC , __SCREAMING_SNAKE_CASE : List[Any] = None , **__SCREAMING_SNAKE_CASE : List[str] , ): A = get_size_dict(_snake_case , default_to_square=_snake_case) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: A = int((2_5_6 / 2_2_4) * size["shortest_edge"]) A = get_resize_output_image_size(_snake_case , size=_snake_case , default_to_square=_snake_case) A = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"""Size dict must have keys \'height\' and \'width\' or \'shortest_edge\'. Got {size_dict.keys()}""") return resize( _snake_case , size=(size_dict["height"], size_dict["width"]) , resample=_snake_case , data_format=_snake_case , **_snake_case) def SCREAMING_SNAKE_CASE__ (self : List[Any] , __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Tuple = None , **__SCREAMING_SNAKE_CASE : int , ): A = get_size_dict(_snake_case) if "height" not in size or "width" not in size: raise ValueError(F"""Size dict must have keys \'height\' and \'width\'. Got {size.keys()}""") return center_crop(_snake_case , size=(size["height"], size["width"]) , data_format=_snake_case , **_snake_case) def SCREAMING_SNAKE_CASE__ (self : Dict , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Tuple = None , **__SCREAMING_SNAKE_CASE : Optional[Any] , ): return rescale(_snake_case , scale=_snake_case , data_format=_snake_case , **_snake_case) def SCREAMING_SNAKE_CASE__ (self : Dict , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : str = None , **__SCREAMING_SNAKE_CASE : Any , ): return normalize(_snake_case , mean=_snake_case , std=_snake_case , data_format=_snake_case , **_snake_case) def SCREAMING_SNAKE_CASE__ (self : Union[str, Any] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[int] = None , __SCREAMING_SNAKE_CASE : int = None , __SCREAMING_SNAKE_CASE : Optional[Any] = None , __SCREAMING_SNAKE_CASE : Optional[Any] = None , __SCREAMING_SNAKE_CASE : List[str] = None , __SCREAMING_SNAKE_CASE : Dict = None , __SCREAMING_SNAKE_CASE : Union[str, Any] = None , __SCREAMING_SNAKE_CASE : Tuple = None , __SCREAMING_SNAKE_CASE : List[str] = None , __SCREAMING_SNAKE_CASE : Tuple = None , __SCREAMING_SNAKE_CASE : int = None , __SCREAMING_SNAKE_CASE : int = ChannelDimension.FIRST , **__SCREAMING_SNAKE_CASE : Any , ): A = do_resize if do_resize is not None else self.do_resize A = resample if resample is not None else self.resample A = do_center_crop if do_center_crop is not None else self.do_center_crop A = do_rescale if do_rescale is not None else self.do_rescale A = rescale_factor if rescale_factor is not None else self.rescale_factor A = do_normalize if do_normalize is not None else self.do_normalize A = image_mean if image_mean is not None else self.image_mean A = image_std if image_std is not None else self.image_std A = size if size is not None else self.size A = get_size_dict(_snake_case , default_to_square=_snake_case) A = crop_size if crop_size is not None else self.crop_size A = get_size_dict(_snake_case , param_name="crop_size") A = make_list_of_images(_snake_case) if not valid_images(_snake_case): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray.") if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True.") if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True.") if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True.") if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True.") # All transformations expect numpy arrays. A = [to_numpy_array(_snake_case) for image in images] if do_resize: A = [self.resize(_snake_case , _snake_case , _snake_case) for image in images] if do_center_crop: A = [self.center_crop(_snake_case , _snake_case) for image in images] if do_rescale: A = [self.rescale(_snake_case , _snake_case) for image in images] if do_normalize: A = [self.normalize(_snake_case , _snake_case , _snake_case) for image in images] A = [to_channel_dimension_format(_snake_case , _snake_case) for image in images] A = {"pixel_values": images} return BatchFeature(data=_snake_case , tensor_type=_snake_case)
350
"""simple docstring""" __A : Dict = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' __A : List[Any] = [{'type': 'code', 'content': INSTALL_CONTENT}] __A : List[Any] = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
57
0
"""simple docstring""" from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING _a = logging.get_logger(__name__) @add_end_docstrings(_lowerCamelCase ) class _lowerCAmelCase ( _lowerCamelCase ): """simple docstring""" def __init__( self : Any, *UpperCAmelCase__ : Union[str, Any], **UpperCAmelCase__ : str ): super().__init__(*_SCREAMING_SNAKE_CASE, **_SCREAMING_SNAKE_CASE ) self.check_model_type(_SCREAMING_SNAKE_CASE ) def _lowercase ( self : str, UpperCAmelCase__ : str=None, UpperCAmelCase__ : str=None, UpperCAmelCase__ : List[str]=None, **UpperCAmelCase__ : Optional[int] ): __lowercase = {}, {} if padding is not None: __lowercase = padding if truncation is not None: __lowercase = truncation if top_k is not None: __lowercase = top_k return preprocess_params, {}, postprocess_params def __call__( self : Tuple, UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Union[str, Any] = None, **UpperCAmelCase__ : List[str] ): if isinstance(_SCREAMING_SNAKE_CASE, (Image.Image, str) ) and isinstance(_SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE ): __lowercase = {'image': image, 'question': question} else: __lowercase = image __lowercase = super().__call__(_SCREAMING_SNAKE_CASE, **_SCREAMING_SNAKE_CASE ) return results def _lowercase ( self : List[str], UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Any=False, UpperCAmelCase__ : str=False ): __lowercase = load_image(inputs["image"] ) __lowercase = self.tokenizer( inputs["question"], return_tensors=self.framework, padding=_SCREAMING_SNAKE_CASE, truncation=_SCREAMING_SNAKE_CASE ) __lowercase = self.image_processor(images=_SCREAMING_SNAKE_CASE, return_tensors=self.framework ) model_inputs.update(_SCREAMING_SNAKE_CASE ) return model_inputs def _lowercase ( self : Optional[Any], UpperCAmelCase__ : List[Any] ): __lowercase = self.model(**_SCREAMING_SNAKE_CASE ) return model_outputs def _lowercase ( self : Optional[int], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Optional[Any]=5 ): if top_k > self.model.config.num_labels: __lowercase = self.model.config.num_labels if self.framework == "pt": __lowercase = model_outputs.logits.sigmoid()[0] __lowercase = probs.topk(_SCREAMING_SNAKE_CASE ) else: raise ValueError(F"""Unsupported framework: {self.framework}""" ) __lowercase = scores.tolist() __lowercase = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(_SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE )]
17
"""simple docstring""" def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : Tuple = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def __lowerCAmelCase (_UpperCamelCase = 100 ): __lowerCAmelCase : Optional[int] = 1 __lowerCAmelCase : Optional[Any] = 2 for i in range(2 , max_n + 1 ): __lowerCAmelCase : Any = pre_numerator __lowerCAmelCase : Union[str, Any] = 2 * i // 3 if i % 3 == 0 else 1 __lowerCAmelCase : int = cur_numerator __lowerCAmelCase : Dict = e_cont * pre_numerator + temp return sum_digits(_UpperCamelCase ) if __name__ == "__main__": print(f'{solution() = }')
86
0
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def snake_case (UpperCAmelCase__ ) -> str: UpperCamelCase_: Dict = FileLock(str(tmpdir / 'foo.lock' ) ) UpperCamelCase_: Tuple = FileLock(str(tmpdir / 'foo.lock' ) ) UpperCamelCase_: List[str] = 0.01 with locka.acquire(): with pytest.raises(UpperCamelCase__ ): UpperCamelCase_: Union[str, Any] = time.time() locka.acquire(UpperCamelCase__ ) assert time.time() - _start > timeout def snake_case (UpperCAmelCase__ ) -> str: UpperCamelCase_: List[Any] = 'a' * 1_0_0_0 + '.lock' UpperCamelCase_: Tuple = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith('.lock' ) assert not locka._lock_file.endswith(UpperCamelCase__ ) assert len(os.path.basename(locka._lock_file ) ) <= 2_5_5 UpperCamelCase_: Any = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(UpperCamelCase__ ): locka.acquire(0 )
363
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def snake_case (UpperCAmelCase__ , UpperCAmelCase__=() , UpperCAmelCase__=None , UpperCAmelCase__="no" , UpperCAmelCase__="29500" ) -> List[Any]: UpperCamelCase_: Any = False UpperCamelCase_: List[str] = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): UpperCamelCase_: List[Any] = True elif "IPython" in sys.modules: UpperCamelCase_: List[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: UpperCamelCase_: Optional[int] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F'''Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.''' ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , UpperCAmelCase__ ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: UpperCamelCase_: List[str] = 8 UpperCamelCase_: str = PrepareForLaunch(UpperCAmelCase__ , distributed_type='TPU' ) print(F'''Launching a training on {num_processes} TPU cores.''' ) xmp.spawn(UpperCAmelCase__ , args=UpperCAmelCase__ , nprocs=UpperCAmelCase__ , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*UpperCAmelCase__ ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=UpperCAmelCase__ , master_addr='127.0.01' , master_port=UpperCAmelCase__ , mixed_precision=UpperCAmelCase__ ): UpperCamelCase_: str = PrepareForLaunch(UpperCAmelCase__ , distributed_type='MULTI_GPU' ) print(F'''Launching training on {num_processes} GPUs.''' ) try: start_processes(UpperCAmelCase__ , args=UpperCAmelCase__ , nprocs=UpperCAmelCase__ , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): UpperCamelCase_: Tuple = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*UpperCAmelCase__ ) def snake_case (UpperCAmelCase__ , UpperCAmelCase__=() , UpperCAmelCase__=2 ) -> Optional[int]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=UpperCAmelCase__ , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): UpperCamelCase_: str = PrepareForLaunch(UpperCAmelCase__ , debug=UpperCAmelCase__ ) start_processes(UpperCAmelCase__ , args=UpperCAmelCase__ , nprocs=UpperCAmelCase__ , start_method='fork' )
292
0
import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class A__ : def __init__( self : Dict , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if dst_width < 0 or dst_height < 0: raise ValueError('Destination width/height should be > 0' ) __lowercase = img __lowercase = img.shape[1] __lowercase = img.shape[0] __lowercase = dst_width __lowercase = dst_height __lowercase = self.src_w / self.dst_w __lowercase = self.src_h / self.dst_h __lowercase = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta ) * 2_55 ) def a__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" for i in range(self.dst_h ): for j in range(self.dst_w ): __lowercase = self.img[self.get_y(_SCREAMING_SNAKE_CASE )][self.get_x(_SCREAMING_SNAKE_CASE )] def a__ ( self : Union[str, Any] , _UpperCAmelCase : int ) -> int: """simple docstring""" return int(self.ratio_x * x ) def a__ ( self : str , _UpperCAmelCase : int ) -> int: """simple docstring""" return int(self.ratio_y * y ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = 800, 600 SCREAMING_SNAKE_CASE__ = imread("""image_data/lena.jpg""", 1) SCREAMING_SNAKE_CASE__ = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( F'''Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}''', n.output ) waitKey(0) destroyAllWindows()
325
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A__ ( unittest.TestCase ): '''simple docstring''' @slow def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small") __lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained("google/mt5-small") __lowerCAmelCase : Tuple = tokenizer("Hello there" , return_tensors="np").input_ids __lowerCAmelCase : Dict = tokenizer("Hi I am" , return_tensors="np").input_ids __lowerCAmelCase : str = shift_tokens_right(_SCREAMING_SNAKE_CASE , model.config.pad_token_id , model.config.decoder_start_token_id) __lowerCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE).logits __lowerCAmelCase : int = optax.softmax_cross_entropy(_SCREAMING_SNAKE_CASE , onehot(_SCREAMING_SNAKE_CASE , logits.shape[-1])).mean() __lowerCAmelCase : List[str] = -(labels.shape[-1] * loss.item()) __lowerCAmelCase : str = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE) < 1e-4)
269
0
'''simple docstring''' from typing import Any def __magic_name__ ( __UpperCAmelCase ) -> list[Any]: '''simple docstring''' if not input_list: return [] snake_case_ = [input_list.count(__UpperCAmelCase ) for value in input_list] snake_case_ = max(__UpperCAmelCase ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(__UpperCAmelCase ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
72
'''simple docstring''' import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class a ( _lowerCamelCase ): snake_case_ = 42 snake_case_ = None def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase=0.9_9_9, __UpperCAmelCase="cosine", ) -> Dict: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(__UpperCAmelCase ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__UpperCAmelCase ): return math.exp(t * -1_2.0 ) else: raise ValueError(F"Unsupported alpha_tranform_type: {alpha_transform_type}" ) snake_case_ = [] for i in range(__UpperCAmelCase ): snake_case_ = i / num_diffusion_timesteps snake_case_ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__UpperCAmelCase ) / alpha_bar_fn(__UpperCAmelCase ), __UpperCAmelCase ) ) return torch.tensor(__UpperCAmelCase, dtype=torch.floataa ) class a ( _lowerCamelCase , _lowerCamelCase ): @register_to_config def __init__( self : List[str] , lowercase_ : int = 1000 , lowercase_ : str = "fixed_small_log" , lowercase_ : bool = True , lowercase_ : Optional[float] = 1.0 , lowercase_ : str = "epsilon" , lowercase_ : str = "squaredcos_cap_v2" , ): if beta_schedule != "squaredcos_cap_v2": raise ValueError('''UnCLIPScheduler only supports `beta_schedule`: \'squaredcos_cap_v2\'''' ) snake_case_ = betas_for_alpha_bar(lowercase_ ) snake_case_ = 1.0 - self.betas snake_case_ = torch.cumprod(self.alphas , dim=0 ) snake_case_ = torch.tensor(1.0 ) # standard deviation of the initial noise distribution snake_case_ = 1.0 # setable values snake_case_ = None snake_case_ = torch.from_numpy(np.arange(0 , lowercase_ )[::-1].copy() ) snake_case_ = variance_type def A_ ( self : Optional[Any] , lowercase_ : torch.FloatTensor , lowercase_ : Optional[int] = None ): return sample def A_ ( self : Optional[int] , lowercase_ : int , lowercase_ : Union[str, torch.device] = None ): snake_case_ = num_inference_steps snake_case_ = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) snake_case_ = (np.arange(0 , lowercase_ ) * step_ratio).round()[::-1].copy().astype(np.intaa ) snake_case_ = torch.from_numpy(lowercase_ ).to(lowercase_ ) def A_ ( self : Optional[int] , lowercase_ : List[Any] , lowercase_ : Optional[int]=None , lowercase_ : Tuple=None , lowercase_ : Tuple=None ): if prev_timestep is None: snake_case_ = t - 1 snake_case_ = self.alphas_cumprod[t] snake_case_ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one snake_case_ = 1 - alpha_prod_t snake_case_ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: snake_case_ = self.betas[t] else: snake_case_ = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample snake_case_ = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: snake_case_ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": snake_case_ = torch.log(torch.clamp(lowercase_ , min=1e-20 ) ) snake_case_ = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler snake_case_ = variance.log() snake_case_ = beta.log() snake_case_ = (predicted_variance + 1) / 2 snake_case_ = frac * max_log + (1 - frac) * min_log return variance def A_ ( self : List[Any] , lowercase_ : torch.FloatTensor , lowercase_ : int , lowercase_ : torch.FloatTensor , lowercase_ : Optional[int] = None , lowercase_ : int=None , lowercase_ : bool = True , ): snake_case_ = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": snake_case_ ,snake_case_ = torch.split(lowercase_ , sample.shape[1] , dim=1 ) else: snake_case_ = None # 1. compute alphas, betas if prev_timestep is None: snake_case_ = t - 1 snake_case_ = self.alphas_cumprod[t] snake_case_ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one snake_case_ = 1 - alpha_prod_t snake_case_ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: snake_case_ = self.betas[t] snake_case_ = self.alphas[t] else: snake_case_ = 1 - alpha_prod_t / alpha_prod_t_prev snake_case_ = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": snake_case_ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": snake_case_ = model_output else: raise ValueError( F"prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`" ''' for the UnCLIPScheduler.''' ) # 3. Clip "predicted x_0" if self.config.clip_sample: snake_case_ = torch.clamp( lowercase_ , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf snake_case_ = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t snake_case_ = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf snake_case_ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise snake_case_ = 0 if t > 0: snake_case_ = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=lowercase_ , device=model_output.device ) snake_case_ = self._get_variance( lowercase_ , predicted_variance=lowercase_ , prev_timestep=lowercase_ , ) if self.variance_type == "fixed_small_log": snake_case_ = variance elif self.variance_type == "learned_range": snake_case_ = (0.5 * variance).exp() else: raise ValueError( F"variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`" ''' for the UnCLIPScheduler.''' ) snake_case_ = variance * variance_noise snake_case_ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=lowercase_ , pred_original_sample=lowercase_ ) def A_ ( self : Any , lowercase_ : torch.FloatTensor , lowercase_ : torch.FloatTensor , lowercase_ : torch.IntTensor , ): # Make sure alphas_cumprod and timestep have same device and dtype as original_samples snake_case_ = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) snake_case_ = timesteps.to(original_samples.device ) snake_case_ = alphas_cumprod[timesteps] ** 0.5 snake_case_ = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): snake_case_ = sqrt_alpha_prod.unsqueeze(-1 ) snake_case_ = (1 - alphas_cumprod[timesteps]) ** 0.5 snake_case_ = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): snake_case_ = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) snake_case_ = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
72
1
'''simple docstring''' from collections import deque def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' A : Dict = len(snake_case__ ) A : str = deque() A : int = [False for _ in range(snake_case__ )] A : List[Any] = [-1 for _ in range(snake_case__ )] A : Optional[int] = index_of[:] def strong_connect(snake_case__ , snake_case__ , snake_case__ ): A : int = index # the number when this node is seen A : Tuple = index # lowest rank node reachable from here index += 1 stack.append(snake_case__ ) A : List[str] = True for w in g[v]: if index_of[w] == -1: A : List[str] = strong_connect(snake_case__ , snake_case__ , snake_case__ ) A : str = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) elif on_stack[w]: A : List[Any] = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) if lowlink_of[v] == index_of[v]: A : Optional[int] = [] A : List[str] = stack.pop() A : List[str] = False component.append(snake_case__ ) while w != v: A : Optional[Any] = stack.pop() A : List[str] = False component.append(snake_case__ ) components.append(snake_case__ ) return index A : Any = [] for v in range(snake_case__ ): if index_of[v] == -1: strong_connect(snake_case__ , 0 , snake_case__ ) return components def lowerCAmelCase_ ( snake_case__ , snake_case__ ): '''simple docstring''' A : Optional[Any] = [[] for _ in range(snake_case__ )] for u, v in edges: g[u].append(snake_case__ ) return g if __name__ == "__main__": # Test lowercase : str = 7 lowercase : Any = [0, 0, 1, 2, 3, 3, 4, 4, 6] lowercase : Tuple = [1, 3, 2, 0, 1, 4, 5, 6, 5] lowercase : Optional[int] = [(u, v) for u, v in zip(source, target)] lowercase : Any = create_graph(n_vertices, edges) assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
3
def a__ ( A_ ): '''simple docstring''' if not isinstance(A_, A_ ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(A_ ) == 0: raise ValueError("""Input list must be a non empty list""" ) if len(A_ ) == 1: return True __magic_name__ = series[1] - series[0] for index in range(len(A_ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def a__ ( A_ ): '''simple docstring''' if not isinstance(A_, A_ ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(A_ ) == 0: raise ValueError("""Input list must be a non empty list""" ) __magic_name__ = 0 for val in series: answer += val return answer / len(A_ ) if __name__ == "__main__": import doctest doctest.testmod()
88
0
'''simple docstring''' import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __lowercase: Dict = get_tests_dir("fixtures/spiece.model") @require_sentencepiece @require_tokenizers class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase): _lowerCamelCase : Union[str, Any] = AlbertTokenizer _lowerCamelCase : str = AlbertTokenizerFast _lowerCamelCase : Tuple = True _lowerCamelCase : List[str] = True _lowerCamelCase : List[str] = True def lowercase_ ( self : Any ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCamelCase__ = AlbertTokenizer(a_ ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ ( self : List[Any], a_ : Optional[Any] ): """simple docstring""" UpperCamelCase__ = "this is a test" UpperCamelCase__ = "this is a test" return input_text, output_text def lowercase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase__ = "<pad>" UpperCamelCase__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ), a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ), a_ ) def lowercase_ ( self : str ): """simple docstring""" UpperCamelCase__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], "<pad>" ) self.assertEqual(vocab_keys[1], "<unk>" ) self.assertEqual(vocab_keys[-1], "▁eloquent" ) self.assertEqual(len(a_ ), 3_0000 ) def lowercase_ ( self : Tuple ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size, 3_0000 ) def lowercase_ ( self : int ): """simple docstring""" if not self.test_rust_tokenizer: return UpperCamelCase__ = self.get_tokenizer() UpperCamelCase__ = self.get_rust_tokenizer() UpperCamelCase__ = "I was born in 92000, and this is falsé." UpperCamelCase__ = tokenizer.tokenize(a_ ) UpperCamelCase__ = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_, a_ ) UpperCamelCase__ = tokenizer.encode(a_, add_special_tokens=a_ ) UpperCamelCase__ = rust_tokenizer.encode(a_, add_special_tokens=a_ ) self.assertListEqual(a_, a_ ) UpperCamelCase__ = self.get_rust_tokenizer() UpperCamelCase__ = tokenizer.encode(a_ ) UpperCamelCase__ = rust_tokenizer.encode(a_ ) self.assertListEqual(a_, a_ ) def lowercase_ ( self : Optional[Any] ): """simple docstring""" UpperCamelCase__ = AlbertTokenizer(a_, keep_accents=a_ ) UpperCamelCase__ = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_, ["▁this", "▁is", "▁a", "▁test"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ), [48, 25, 21, 1289] ) UpperCamelCase__ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( a_, ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", "."] ) UpperCamelCase__ = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual(a_, [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) UpperCamelCase__ = tokenizer.convert_ids_to_tokens(a_ ) self.assertListEqual( a_, ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "."], ) def lowercase_ ( self : List[str] ): """simple docstring""" UpperCamelCase__ = AlbertTokenizer(a_ ) UpperCamelCase__ = tokenizer.encode("sequence builders" ) UpperCamelCase__ = tokenizer.encode("multi-sequence build" ) UpperCamelCase__ = tokenizer.build_inputs_with_special_tokens(a_ ) UpperCamelCase__ = tokenizer.build_inputs_with_special_tokens(a_, a_ ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def lowercase_ ( self : Tuple ): """simple docstring""" UpperCamelCase__ = {"attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "input_ids": [[2, 2_1970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 1_2051, 18, 17, 7103, 2153, 673, 8, 3515, 1_8684, 8, 4461, 6, 1927, 297, 8, 1_2060, 2607, 18, 13, 5, 4461, 15, 1_0538, 38, 8, 135, 15, 822, 58, 15, 993, 1_0363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 1_0641, 6, 29, 84, 2512, 2430, 782, 1_8684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 1_1712, 15, 7103, 2153, 673, 17, 2_4883, 9990, 9, 3], [2, 1_1502, 25, 1006, 20, 782, 8, 1_1809, 855, 1732, 1_9393, 1_8667, 37, 367, 2_1018, 69, 1854, 34, 1_1860, 1_9124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 1_7659, 84, 14, 1_6792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_, model_name="albert-base-v2", revision="6b6560eaf5ff2e250b00c50f380c5389a9c2d82e", )
367
'''simple docstring''' def SCREAMING_SNAKE_CASE__( _UpperCamelCase : str ) -> bool: '''simple docstring''' return credit_card_number.startswith(("34", "35", "37", "4", "5", "6") ) def SCREAMING_SNAKE_CASE__( _UpperCamelCase : str ) -> bool: '''simple docstring''' UpperCamelCase__ = credit_card_number UpperCamelCase__ = 0 UpperCamelCase__ = len(_UpperCamelCase ) - 2 for i in range(_UpperCamelCase , -1 , -2 ): # double the value of every second digit UpperCamelCase__ = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 UpperCamelCase__ = cc_number[:i] + str(_UpperCamelCase ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(_UpperCamelCase ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def SCREAMING_SNAKE_CASE__( _UpperCamelCase : str ) -> bool: '''simple docstring''' UpperCamelCase__ = F'{credit_card_number} is an invalid credit card number because' if not credit_card_number.isdigit(): print(F'{error_message} it has nonnumerical characters.' ) return False if not 13 <= len(_UpperCamelCase ) <= 16: print(F'{error_message} of its length.' ) return False if not validate_initial_digits(_UpperCamelCase ): print(F'{error_message} of its first two digits.' ) return False if not luhn_validation(_UpperCamelCase ): print(F'{error_message} it fails the Luhn check.' ) return False print(F'{credit_card_number} is a valid credit card number.' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number("4111111111111111") validate_credit_card_number("32323")
31
0
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase_ ( __A, __A=False ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("head" ): UpperCAmelCase__ = "segformer.encoder." + key if key.startswith("backbone" ): UpperCAmelCase__ = key.replace("backbone", "segformer.encoder" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 UpperCAmelCase__ = key[key.find("patch_embed" ) + len("patch_embed" )] UpperCAmelCase__ = key.replace(f"""patch_embed{idx}""", f"""patch_embeddings.{int(__A )-1}""" ) if "norm" in key: UpperCAmelCase__ = key.replace("norm", "layer_norm" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 UpperCAmelCase__ = key[key.find("segformer.encoder.layer_norm" ) + len("segformer.encoder.layer_norm" )] UpperCAmelCase__ = key.replace(f"""layer_norm{idx}""", f"""layer_norm.{int(__A )-1}""" ) if "layer_norm1" in key: UpperCAmelCase__ = key.replace("layer_norm1", "layer_norm_1" ) if "layer_norm2" in key: UpperCAmelCase__ = key.replace("layer_norm2", "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 UpperCAmelCase__ = key[key.find("block" ) + len("block" )] UpperCAmelCase__ = key.replace(f"""block{idx}""", f"""block.{int(__A )-1}""" ) if "attn.q" in key: UpperCAmelCase__ = key.replace("attn.q", "attention.self.query" ) if "attn.proj" in key: UpperCAmelCase__ = key.replace("attn.proj", "attention.output.dense" ) if "attn" in key: UpperCAmelCase__ = key.replace("attn", "attention.self" ) if "fc1" in key: UpperCAmelCase__ = key.replace("fc1", "dense1" ) if "fc2" in key: UpperCAmelCase__ = key.replace("fc2", "dense2" ) if "linear_pred" in key: UpperCAmelCase__ = key.replace("linear_pred", "classifier" ) if "linear_fuse" in key: UpperCAmelCase__ = key.replace("linear_fuse.conv", "linear_fuse" ) UpperCAmelCase__ = key.replace("linear_fuse.bn", "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 UpperCAmelCase__ = key[key.find("linear_c" ) + len("linear_c" )] UpperCAmelCase__ = key.replace(f"""linear_c{idx}""", f"""linear_c.{int(__A )-1}""" ) if key.startswith("head" ): UpperCAmelCase__ = key.replace("head", "classifier" ) UpperCAmelCase__ = value return new_state_dict def lowerCAmelCase_ ( __A, __A ) -> Any: '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) UpperCAmelCase__ = state_dict.pop(f"""segformer.encoder.block.{i}.{j}.attention.self.kv.weight""" ) UpperCAmelCase__ = state_dict.pop(f"""segformer.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict UpperCAmelCase__ = kv_weight[ : config.hidden_sizes[i], : ] UpperCAmelCase__ = kv_bias[: config.hidden_sizes[i]] UpperCAmelCase__ = kv_weight[ config.hidden_sizes[i] :, : ] UpperCAmelCase__ = kv_bias[ config.hidden_sizes[i] : ] def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase__ = Image.open(requests.get(__A, stream=__A ).raw ) return image @torch.no_grad() def lowerCAmelCase_ ( __A, __A, __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = SegformerConfig() UpperCAmelCase__ = False # set attributes based on model_name UpperCAmelCase__ = "huggingface/label-files" if "segformer" in model_name: UpperCAmelCase__ = model_name[len("segformer." ) : len("segformer." ) + 2] if "ade" in model_name: UpperCAmelCase__ = 150 UpperCAmelCase__ = "ade20k-id2label.json" UpperCAmelCase__ = (1, 150, 128, 128) elif "city" in model_name: UpperCAmelCase__ = 19 UpperCAmelCase__ = "cityscapes-id2label.json" UpperCAmelCase__ = (1, 19, 128, 128) else: raise ValueError(f"""Model {model_name} not supported""" ) elif "mit" in model_name: UpperCAmelCase__ = True UpperCAmelCase__ = model_name[4:6] UpperCAmelCase__ = 1_000 UpperCAmelCase__ = "imagenet-1k-id2label.json" UpperCAmelCase__ = (1, 1_000) else: raise ValueError(f"""Model {model_name} not supported""" ) # set config attributes UpperCAmelCase__ = json.load(open(hf_hub_download(__A, __A, repo_type="dataset" ), "r" ) ) UpperCAmelCase__ = {int(__A ): v for k, v in idalabel.items()} UpperCAmelCase__ = idalabel UpperCAmelCase__ = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": UpperCAmelCase__ = [64, 128, 320, 512] UpperCAmelCase__ = 256 elif size == "b2": UpperCAmelCase__ = [64, 128, 320, 512] UpperCAmelCase__ = 768 UpperCAmelCase__ = [3, 4, 6, 3] elif size == "b3": UpperCAmelCase__ = [64, 128, 320, 512] UpperCAmelCase__ = 768 UpperCAmelCase__ = [3, 4, 18, 3] elif size == "b4": UpperCAmelCase__ = [64, 128, 320, 512] UpperCAmelCase__ = 768 UpperCAmelCase__ = [3, 8, 27, 3] elif size == "b5": UpperCAmelCase__ = [64, 128, 320, 512] UpperCAmelCase__ = 768 UpperCAmelCase__ = [3, 6, 40, 3] else: raise ValueError(f"""Size {size} not supported""" ) # load image processor (only resize + normalize) UpperCAmelCase__ = SegformerImageProcessor( image_scale=(512, 512), keep_ratio=__A, align=__A, do_random_crop=__A ) # prepare image UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=__A, return_tensors="pt" ).pixel_values logger.info(f"""Converting model {model_name}...""" ) # load original state dict if encoder_only: UpperCAmelCase__ = torch.load(__A, map_location=torch.device("cpu" ) ) else: UpperCAmelCase__ = torch.load(__A, map_location=torch.device("cpu" ) )["state_dict"] # rename keys UpperCAmelCase__ = rename_keys(__A, encoder_only=__A ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(__A, __A ) # create HuggingFace model and load state dict if encoder_only: UpperCAmelCase__ = False UpperCAmelCase__ = SegformerForImageClassification(__A ) else: UpperCAmelCase__ = SegformerForSemanticSegmentation(__A ) model.load_state_dict(__A ) model.eval() # forward pass UpperCAmelCase__ = model(__A ) UpperCAmelCase__ = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": UpperCAmelCase__ = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": UpperCAmelCase__ = torch.tensor( [ [[-7.5820, -8.7231, -8.3215], [-8.0600, -10.3529, -10.0304], [-7.5208, -9.4103, -9.6239]], [[-12.6918, -13.8994, -13.7137], [-13.3196, -15.7523, -15.4789], [-12.9343, -14.8757, -14.9689]], [[-11.1911, -11.9421, -11.3243], [-11.3342, -13.6839, -13.3581], [-10.3909, -12.1832, -12.4858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": UpperCAmelCase__ = torch.tensor( [ [[-11.8173, -14.3850, -16.3128], [-14.5648, -16.5804, -18.6568], [-14.7223, -15.7387, -18.4218]], [[-15.7290, -17.9171, -19.4423], [-18.3105, -19.9448, -21.4661], [-17.9296, -18.6497, -20.7910]], [[-15.0783, -17.0336, -18.2789], [-16.8771, -18.6870, -20.1612], [-16.2454, -17.1426, -19.5055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": UpperCAmelCase__ = torch.tensor( [ [[-9.0878, -10.2081, -10.1891], [-9.3144, -10.7941, -10.9843], [-9.2294, -10.3855, -10.5704]], [[-12.2316, -13.9068, -13.6102], [-12.9161, -14.3702, -14.3235], [-12.5233, -13.7174, -13.7932]], [[-14.6275, -15.2490, -14.9727], [-14.3400, -15.9687, -16.2827], [-14.1484, -15.4033, -15.8937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": UpperCAmelCase__ = torch.tensor( [ [[-12.3144, -13.2447, -14.0802], [-13.3614, -14.5816, -15.6117], [-13.3340, -14.4433, -16.2219]], [[-19.2781, -20.4128, -20.7506], [-20.6153, -21.6566, -22.0998], [-19.9800, -21.0430, -22.1494]], [[-18.8739, -19.7804, -21.1834], [-20.1233, -21.6765, -23.2944], [-20.0315, -21.2641, -23.6944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": UpperCAmelCase__ = torch.tensor( [ [[-9.5524, -12.0835, -11.7348], [-10.5229, -13.6446, -14.5662], [-9.5842, -12.8851, -13.9414]], [[-15.3432, -17.5323, -17.0818], [-16.3330, -18.9255, -19.2101], [-15.1340, -17.7848, -18.3971]], [[-12.6072, -14.9486, -14.6631], [-13.7629, -17.0907, -17.7745], [-12.7899, -16.1695, -17.1671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-11.9295, -13.4057, -14.8106], [-13.3431, -14.8179, -15.3781], [-14.2836, -15.5942, -16.1588]], [[-11.4906, -12.8067, -13.6564], [-13.1189, -14.0500, -14.1543], [-13.8748, -14.5136, -14.8789]], [[0.5374, 0.1067, -0.4742], [0.1141, -0.2255, -0.7099], [-0.3000, -0.5924, -1.3105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-7.8217, -9.8767, -10.1717], [-9.4438, -10.9058, -11.4047], [-9.7939, -12.3495, -12.1079]], [[-7.1514, -9.5336, -10.0860], [-9.7776, -11.6822, -11.8439], [-10.1411, -12.7655, -12.8972]], [[0.3021, 0.0805, -0.2310], [-0.0328, -0.1605, -0.2714], [-0.1408, -0.5477, -0.6976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": UpperCAmelCase__ = torch.tensor( [ [ [-1.1_372e01, -1.2_787e01, -1.3_477e01], [-1.2_536e01, -1.4_194e01, -1.4_409e01], [-1.3_217e01, -1.4_888e01, -1.5_327e01], ], [ [-1.4_791e01, -1.7_122e01, -1.8_277e01], [-1.7_163e01, -1.9_192e01, -1.9_533e01], [-1.7_897e01, -1.9_991e01, -2.0_315e01], ], [ [7.6_723e-01, 4.1_921e-01, -7.7_878e-02], [4.7_772e-01, 9.5_557e-03, -2.8_082e-01], [3.6_032e-01, -2.4_826e-01, -5.1_168e-01], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-9.4959, -11.3087, -11.7479], [-11.0025, -12.6540, -12.3319], [-11.4064, -13.0487, -12.9905]], [[-9.8905, -11.3084, -12.0854], [-11.1726, -12.7698, -12.9583], [-11.5985, -13.3278, -14.1774]], [[0.2213, 0.0192, -0.2466], [-0.1731, -0.4213, -0.4874], [-0.3126, -0.6541, -1.1389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-16.0976, -16.4856, -17.3962], [-16.6234, -19.0342, -19.7685], [-16.0900, -18.0661, -19.1180]], [[-18.4750, -18.8488, -19.5074], [-19.4030, -22.1570, -22.5977], [-19.1191, -20.8486, -22.3783]], [[-4.5178, -5.5037, -6.5109], [-5.0884, -7.2174, -8.0334], [-4.4156, -5.8117, -7.2970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-14.2081, -14.4732, -14.1977], [-14.5867, -16.4423, -16.6356], [-13.4441, -14.9685, -16.8696]], [[-14.4576, -14.7073, -15.0451], [-15.0816, -17.6237, -17.9873], [-14.4213, -16.0199, -18.5992]], [[-4.7349, -4.9588, -5.0966], [-4.3210, -6.9325, -7.2591], [-3.4312, -4.7484, -7.1917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-11.7737, -11.9526, -11.3273], [-13.6692, -14.4574, -13.8878], [-13.8937, -14.6924, -15.9345]], [[-14.6706, -14.5330, -14.1306], [-16.1502, -16.8180, -16.4269], [-16.8338, -17.8939, -20.1746]], [[1.0491, 0.8289, 1.0310], [1.1044, 0.5219, 0.8055], [1.0899, 0.6926, 0.5590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": UpperCAmelCase__ = torch.tensor( [ [[-12.5641, -13.4777, -13.0684], [-13.9587, -15.8983, -16.6557], [-13.3109, -15.7350, -16.3141]], [[-14.7074, -15.4352, -14.5944], [-16.6353, -18.1663, -18.6120], [-15.1702, -18.0329, -18.1547]], [[-1.7990, -2.0951, -1.7784], [-2.6397, -3.8245, -3.9686], [-1.5264, -2.8126, -2.9316]], ] ) else: UpperCAmelCase__ = logits.argmax(-1 ).item() print("Predicted class:", model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3], __A, atol=1e-2 ) # finally, save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(__A ).mkdir(exist_ok=__A ) model.save_pretrained(__A ) image_processor.save_pretrained(__A ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument( '--model_name', default='segformer.b0.512x512.ade.160k', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) UpperCamelCase__ = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
65
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING, TextaTextGenerationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, require_tf, require_torch from transformers.utils import is_torch_available from .test_pipelines_common import ANY if is_torch_available(): import torch @is_pipeline_test class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] =MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __UpperCAmelCase : Union[str, Any] =TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING def snake_case ( self , __a , __a , __a ): __lowerCAmelCase = TextaTextGenerationPipeline(model=__a , tokenizer=__a ) return generator, ["Something to write", "Something else"] def snake_case ( self , __a , __a ): __lowerCAmelCase = generator("Something there" ) self.assertEqual(__a , [{"generated_text": ANY(__a )}] ) # These are encoder decoder, they don't just append to incoming string self.assertFalse(outputs[0]["generated_text"].startswith("Something there" ) ) __lowerCAmelCase = generator(["This is great !", "Something else"] , num_return_sequences=2 , do_sample=__a ) self.assertEqual( __a , [ [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], ] , ) __lowerCAmelCase = generator( ["This is great !", "Something else"] , num_return_sequences=2 , batch_size=2 , do_sample=__a ) self.assertEqual( __a , [ [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], [{"generated_text": ANY(__a )}, {"generated_text": ANY(__a )}], ] , ) with self.assertRaises(__a ): generator(4 ) @require_torch def snake_case ( self ): __lowerCAmelCase = pipeline("text2text-generation" , model="patrickvonplaten/t5-tiny-random" , framework="pt" ) # do_sample=False necessary for reproducibility __lowerCAmelCase = generator("Something there" , do_sample=__a ) self.assertEqual(__a , [{"generated_text": ""}] ) __lowerCAmelCase = 3 __lowerCAmelCase = generator( "Something there" , num_return_sequences=__a , num_beams=__a , ) __lowerCAmelCase = [ {"generated_text": "Beide Beide Beide Beide Beide Beide Beide Beide Beide"}, {"generated_text": "Beide Beide Beide Beide Beide Beide Beide Beide"}, {"generated_text": ""}, ] self.assertEqual(__a , __a ) __lowerCAmelCase = generator("This is a test" , do_sample=__a , num_return_sequences=2 , return_tensors=__a ) self.assertEqual( __a , [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ] , ) __lowerCAmelCase = generator.model.config.eos_token_id __lowerCAmelCase = "<pad>" __lowerCAmelCase = generator( ["This is a test", "This is a second test"] , do_sample=__a , num_return_sequences=2 , batch_size=2 , return_tensors=__a , ) self.assertEqual( __a , [ [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ], [ {"generated_token_ids": ANY(torch.Tensor )}, {"generated_token_ids": ANY(torch.Tensor )}, ], ] , ) @require_tf def snake_case ( self ): __lowerCAmelCase = pipeline("text2text-generation" , model="patrickvonplaten/t5-tiny-random" , framework="tf" ) # do_sample=False necessary for reproducibility __lowerCAmelCase = generator("Something there" , do_sample=__a ) self.assertEqual(__a , [{"generated_text": ""}] )
57
0
'''simple docstring''' def __lowerCamelCase ( A__ , A__ ) -> float: """simple docstring""" if density <= 0: raise ValueError('Impossible fluid density' ) if bulk_modulus <= 0: raise ValueError('Impossible bulk modulus' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
355
'''simple docstring''' import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = (UnCLIPScheduler,) def A ( self : Union[str, Any] , **UpperCamelCase__ : Any ): """simple docstring""" UpperCamelCase = { 'num_train_timesteps': 1_0_0_0, 'variance_type': 'fixed_small_log', 'clip_sample': True, 'clip_sample_range': 1.0, 'prediction_type': 'epsilon', } config.update(**UpperCamelCase__ ) return config def A ( self : str ): """simple docstring""" for timesteps in [1, 5, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=UpperCamelCase__ ) def A ( self : List[str] ): """simple docstring""" for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=UpperCamelCase__ ) def A ( self : List[Any] ): """simple docstring""" for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCamelCase__ ) def A ( self : Optional[int] ): """simple docstring""" for clip_sample_range in [1, 5, 1_0, 2_0]: self.check_over_configs(clip_sample_range=UpperCamelCase__ ) def A ( self : Tuple ): """simple docstring""" for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=UpperCamelCase__ ) def A ( self : Union[str, Any] ): """simple docstring""" for time_step in [0, 5_0_0, 9_9_9]: for prev_timestep in [None, 5, 1_0_0, 2_5_0, 5_0_0, 7_5_0]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=UpperCamelCase__ , prev_timestep=UpperCamelCase__ ) def A ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config(variance_type='fixed_small_log' ) UpperCamelCase = scheduler_class(**UpperCamelCase__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0_0_0_0E-1_0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_8_7 ) - 0.0_5_4_9_6_2_5 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_9_9 ) - 0.9_9_9_4_9_8_7 ) ) < 1E-5 def A ( self : Tuple ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config(variance_type='learned_range' ) UpperCamelCase = scheduler_class(**UpperCamelCase__ ) UpperCamelCase = 0.5 assert scheduler._get_variance(1 , predicted_variance=UpperCamelCase__ ) - -1_0.1_7_1_2_7_9_0 < 1E-5 assert scheduler._get_variance(4_8_7 , predicted_variance=UpperCamelCase__ ) - -5.7_9_9_8_0_5_2 < 1E-5 assert scheduler._get_variance(9_9_9 , predicted_variance=UpperCamelCase__ ) - -0.0_0_1_0_0_1_1 < 1E-5 def A ( self : int ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config() UpperCamelCase = scheduler_class(**UpperCamelCase__ ) UpperCamelCase = scheduler.timesteps UpperCamelCase = self.dummy_model() UpperCamelCase = self.dummy_sample_deter UpperCamelCase = torch.manual_seed(0 ) for i, t in enumerate(UpperCamelCase__ ): # 1. predict noise residual UpperCamelCase = model(UpperCamelCase__ , UpperCamelCase__ ) # 2. predict previous mean of sample x_t-1 UpperCamelCase = scheduler.step(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample UpperCamelCase = pred_prev_sample UpperCamelCase = torch.sum(torch.abs(UpperCamelCase__ ) ) UpperCamelCase = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 2_5_2.2_6_8_2_4_9_5 ) < 1E-2 assert abs(result_mean.item() - 0.3_2_8_4_7_4_3 ) < 1E-3 def A ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config() UpperCamelCase = scheduler_class(**UpperCamelCase__ ) scheduler.set_timesteps(2_5 ) UpperCamelCase = scheduler.timesteps UpperCamelCase = self.dummy_model() UpperCamelCase = self.dummy_sample_deter UpperCamelCase = torch.manual_seed(0 ) for i, t in enumerate(UpperCamelCase__ ): # 1. predict noise residual UpperCamelCase = model(UpperCamelCase__ , UpperCamelCase__ ) if i + 1 == timesteps.shape[0]: UpperCamelCase = None else: UpperCamelCase = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 UpperCamelCase = scheduler.step( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , prev_timestep=UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample UpperCamelCase = pred_prev_sample UpperCamelCase = torch.sum(torch.abs(UpperCamelCase__ ) ) UpperCamelCase = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 2_5_8.2_0_4_4_9_8_3 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_6_2_0_3_8 ) < 1E-3 def A ( self : Tuple ): """simple docstring""" pass def A ( self : Optional[int] ): """simple docstring""" pass
249
0
'''simple docstring''' from decimal import Decimal, getcontext from math import ceil, factorial def __lowerCamelCase ( A__ ) -> str: """simple docstring""" if not isinstance(A__ , A__ ): raise TypeError('Undefined for non-integers' ) elif precision < 1: raise ValueError('Undefined for non-natural numbers' ) UpperCamelCase = precision UpperCamelCase = ceil(precision / 14 ) UpperCamelCase = 426_880 * Decimal(10_005 ).sqrt() UpperCamelCase = 1 UpperCamelCase = 13_591_409 UpperCamelCase = Decimal(A__ ) for k in range(1 , A__ ): UpperCamelCase = factorial(6 * k ) // (factorial(3 * k ) * factorial(A__ ) ** 3) linear_term += 545_140_134 exponential_term *= -262_537_412_640_768_000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": _lowerCamelCase : Optional[int] = 50 print(f'''The first {n} digits of pi is: {pi(n)}''')
28
"""simple docstring""" from __future__ import annotations import unittest from transformers import AutoTokenizer, PegasusConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFPegasusForConditionalGeneration, TFPegasusModel @require_tf class _UpperCAmelCase : UpperCamelCase = PegasusConfig UpperCamelCase = {} UpperCamelCase = '''gelu''' def __init__( self :Union[str, Any] , __UpperCamelCase :Union[str, Any] , __UpperCamelCase :str=13 , __UpperCamelCase :List[Any]=7 , __UpperCamelCase :Union[str, Any]=True , __UpperCamelCase :List[Any]=False , __UpperCamelCase :Any=99 , __UpperCamelCase :Tuple=32 , __UpperCamelCase :Optional[int]=2 , __UpperCamelCase :Optional[Any]=4 , __UpperCamelCase :Tuple=37 , __UpperCamelCase :Optional[Any]=0.1 , __UpperCamelCase :Tuple=0.1 , __UpperCamelCase :Optional[int]=40 , __UpperCamelCase :Tuple=2 , __UpperCamelCase :Dict=1 , __UpperCamelCase :Any=0 , ): A = parent A = batch_size A = seq_length A = is_training A = use_labels A = vocab_size A = hidden_size A = num_hidden_layers A = num_attention_heads A = intermediate_size A = hidden_dropout_prob A = attention_probs_dropout_prob A = max_position_embeddings A = eos_token_id A = pad_token_id A = bos_token_id def lowerCamelCase ( self :Tuple ): A = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) A = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) A = tf.concat([input_ids, eos_tensor] , axis=1 ) A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) A = prepare_pegasus_inputs_dict(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return config, inputs_dict def lowerCamelCase ( self :str , __UpperCamelCase :str , __UpperCamelCase :Union[str, Any] ): A = TFPegasusModel(config=__UpperCamelCase ).get_decoder() A = inputs_dict["input_ids"] A = input_ids[:1, :] A = inputs_dict["attention_mask"][:1, :] A = inputs_dict["head_mask"] A = 1 # first forward pass A = model(__UpperCamelCase , attention_mask=__UpperCamelCase , head_mask=__UpperCamelCase , use_cache=__UpperCamelCase ) A, A = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids A = ids_tensor((self.batch_size, 3) , config.vocab_size ) A = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and A = tf.concat([input_ids, next_tokens] , axis=-1 ) A = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) A = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] A = model(__UpperCamelCase , attention_mask=__UpperCamelCase , past_key_values=__UpperCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice A = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) A = output_from_no_past[:, -3:, random_slice_idx] A = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1e-3 ) def A__ ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase=None , ): if attention_mask is None: A = tf.cast(tf.math.not_equal(UpperCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: A = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: A = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: A = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: A = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _UpperCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): UpperCamelCase = (TFPegasusForConditionalGeneration, TFPegasusModel) if is_tf_available() else () UpperCamelCase = (TFPegasusForConditionalGeneration,) if is_tf_available() else () UpperCamelCase = ( { '''conversational''': TFPegasusForConditionalGeneration, '''feature-extraction''': TFPegasusModel, '''summarization''': TFPegasusForConditionalGeneration, '''text2text-generation''': TFPegasusForConditionalGeneration, '''translation''': TFPegasusForConditionalGeneration, } if is_tf_available() else {} ) UpperCamelCase = True UpperCamelCase = False UpperCamelCase = False def lowerCamelCase ( self :int ): A = TFPegasusModelTester(self ) A = ConfigTester(self , config_class=__UpperCamelCase ) def lowerCamelCase ( self :Dict ): self.config_tester.run_common_tests() def lowerCamelCase ( self :Any ): A = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCamelCase ) @require_sentencepiece @require_tokenizers @require_tf class _UpperCAmelCase ( unittest.TestCase ): UpperCamelCase = [ ''' PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.''', ''' The London trio are up for best UK act and best album, as well as getting two nominations in the best song category."We got told like this morning \'Oh I think you\'re nominated\'", said Dappy."And I was like \'Oh yeah, which one?\' And now we\'ve got nominated for four awards. I mean, wow!"Bandmate Fazer added: "We thought it\'s best of us to come down and mingle with everyone and say hello to the cameras. And now we find we\'ve got four nominations."The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn\'t be too disappointed if they didn\'t win this time around."At the end of the day we\'re grateful to be where we are in our careers."If it don\'t happen then it don\'t happen - live to fight another day and keep on making albums and hits for the fans."Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers\' All These Things That I\'ve Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year\'s Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border."We just done Edinburgh the other day," said Dappy."We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!" ''', ] UpperCamelCase = [ '''California\'s largest electricity provider has cut power to hundreds of thousands of customers in an effort to''' ''' reduce the risk of wildfires.''', '''N-Dubz have revealed they\'re "grateful" to have been nominated for four Mobo Awards.''', ] # differs slightly from pytorch, likely due to numerical differences in linear layers UpperCamelCase = '''google/pegasus-xsum''' @cached_property def lowerCamelCase ( self :Any ): return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def lowerCamelCase ( self :Dict ): A = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def lowerCamelCase ( self :str , **__UpperCamelCase :str ): A = self.translate_src_text(**__UpperCamelCase ) assert self.expected_text == generated_words def lowerCamelCase ( self :Any , **__UpperCamelCase :List[str] ): A = self.tokenizer(self.src_text , **__UpperCamelCase , padding=__UpperCamelCase , return_tensors="tf" ) A = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=__UpperCamelCase , ) A = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=__UpperCamelCase ) return generated_words @slow def lowerCamelCase ( self :Union[str, Any] ): self._assert_generated_batch_equal_expected()
292
0
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class UpperCAmelCase_ : '''simple docstring''' def __init__( self , _lowercase ): """simple docstring""" _lowerCAmelCase = str(id_ ) _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = [] _lowerCAmelCase = {} # {vertex:distance} def __lt__( self , _lowercase ): """simple docstring""" return self.key < other.key def __repr__( self ): """simple docstring""" return self.id def _lowercase ( self , _lowercase ): """simple docstring""" self.neighbors.append(_lowercase ) def _lowercase ( self , _lowercase , _lowercase ): """simple docstring""" _lowerCAmelCase = weight def A (__lowerCamelCase :List[Any] , __lowerCamelCase :Union[str, Any] , __lowerCamelCase :Dict , __lowerCamelCase :Optional[int] ): # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __lowerCamelCase ) graph[b - 1].add_edge(graph[a - 1] , __lowerCamelCase ) def A (__lowerCamelCase :list , __lowerCamelCase :Vertex ): _lowerCAmelCase = [] for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = graph[:] while q: _lowerCAmelCase = min(__lowerCamelCase ) q.remove(__lowerCamelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] for i in range(1 , len(__lowerCamelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def A (__lowerCamelCase :list , __lowerCamelCase :Vertex ): for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = list(__lowerCamelCase ) hq.heapify(__lowerCamelCase ) while h: _lowerCAmelCase = hq.heappop(__lowerCamelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] hq.heapify(__lowerCamelCase ) for i in range(1 , len(__lowerCamelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def A (): pass if __name__ == "__main__": import doctest doctest.testmod()
229
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _lowercase = datasets.logging.get_logger(__name__) _lowercase = """\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\", author = \"Moosavi, Nafise Sadat and Strube, Michael\", booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\", month = aug, year = \"2016\", address = \"Berlin, Germany\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/P16-1060\", doi = \"10.18653/v1/P16-1060\", pages = \"632--642\", } """ _lowercase = """\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. """ _lowercase = """ Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting 'keep_singletons=False', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs. min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: 'mentions': mentions 'muc': MUC metric [Vilain et al, 1995] 'bcub': B-cubed [Bagga and Baldwin, 1998] 'ceafe': CEAFe [Luo et al., 2005] 'lea': LEA [Moosavi and Strube, 2016] 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric('coval') >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -', ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)', ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)', ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -', ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -', ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {'mentions/recall': 1.0,[...] 'conll_score': 100.0} """ def A (__lowerCamelCase :str , __lowerCamelCase :Optional[Any] , __lowerCamelCase :Union[str, Any]=False , __lowerCamelCase :List[Any]=False , __lowerCamelCase :str=True , __lowerCamelCase :str=False , __lowerCamelCase :str="dummy_doc" ): _lowerCAmelCase = {doc: key_lines} _lowerCAmelCase = {doc: sys_lines} _lowerCAmelCase = {} _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase , _lowerCAmelCase = reader.get_doc_mentions(__lowerCamelCase , key_doc_lines[doc] , __lowerCamelCase ) key_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase = reader.set_annotated_parse_trees(__lowerCamelCase , key_doc_lines[doc] , __lowerCamelCase , __lowerCamelCase ) _lowerCAmelCase , _lowerCAmelCase = reader.get_doc_mentions(__lowerCamelCase , sys_doc_lines[doc] , __lowerCamelCase ) sys_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase = reader.set_annotated_parse_trees(__lowerCamelCase , key_doc_lines[doc] , __lowerCamelCase , __lowerCamelCase ) if remove_nested: _lowerCAmelCase , _lowerCAmelCase = reader.remove_nested_coref_mentions(__lowerCamelCase , __lowerCamelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _lowerCAmelCase , _lowerCAmelCase = reader.remove_nested_coref_mentions(__lowerCamelCase , __lowerCamelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _lowerCAmelCase = reader.get_mention_assignments(__lowerCamelCase , __lowerCamelCase ) _lowerCAmelCase = reader.get_mention_assignments(__lowerCamelCase , __lowerCamelCase ) _lowerCAmelCase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( """Number of resulting singleton clusters in the key """ f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' """files, respectively""" ) return doc_coref_infos def A (__lowerCamelCase :List[str] , __lowerCamelCase :str , __lowerCamelCase :str , __lowerCamelCase :int , __lowerCamelCase :int , __lowerCamelCase :Optional[Any] , __lowerCamelCase :Optional[Any] ): _lowerCAmelCase = get_coref_infos(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) _lowerCAmelCase = {} _lowerCAmelCase = 0 _lowerCAmelCase = 0 for name, metric in metrics: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = evaluator.evaluate_documents(__lowerCamelCase , __lowerCamelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(10 ) , f'Recall: {recall * 100:.2f}' , f' Precision: {precision * 100:.2f}' , f' F1: {fa * 100:.2f}' , ) if conll_subparts_num == 3: _lowerCAmelCase = (conll / 3) * 100 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({"""conll_score""": conll} ) return output_scores def A (__lowerCamelCase :List[str] ): _lowerCAmelCase = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: _lowerCAmelCase = line.split()[5] if not parse_col == "-": _lowerCAmelCase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase_ ( datasets.Metric ): '''simple docstring''' def _lowercase ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def _lowercase ( self , _lowercase , _lowercase , _lowercase=True , _lowercase=False , _lowercase=False , _lowercase=False ): """simple docstring""" _lowerCAmelCase = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _lowerCAmelCase = util.check_gold_parse_annotation(_lowercase ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _lowerCAmelCase = evaluate( key_lines=_lowercase , sys_lines=_lowercase , metrics=_lowercase , NP_only=_lowercase , remove_nested=_lowercase , keep_singletons=_lowercase , min_span=_lowercase , ) return score
229
1
"""simple docstring""" # limitations under the License. from typing import Optional, Tuple, Union import torch from diffusers import DiffusionPipeline, ImagePipelineOutput class __snake_case ( _lowercase): def __init__( self : Union[str, Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Dict ): """simple docstring""" super().__init__() self.register_modules(unet=__lowerCAmelCase , scheduler=__lowerCAmelCase ) @torch.no_grad() def __call__( self : List[Any] , __lowerCAmelCase : int = 1 , __lowerCAmelCase : Optional[torch.Generator] = None , __lowerCAmelCase : int = 5_0 , __lowerCAmelCase : Optional[str] = "pil" , __lowerCAmelCase : bool = True , **__lowerCAmelCase : Tuple , ): """simple docstring""" _lowerCamelCase : int = torch.randn( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=__lowerCAmelCase , ) _lowerCamelCase : Union[str, Any] = image.to(self.device ) # set step values self.scheduler.set_timesteps(__lowerCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output _lowerCamelCase : List[str] = self.unet(__lowerCAmelCase , __lowerCAmelCase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _lowerCamelCase : List[str] = self.scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ).prev_sample _lowerCamelCase : List[str] = (image / 2 + 0.5).clamp(0 , 1 ) _lowerCamelCase : int = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _lowerCamelCase : List[str] = self.numpy_to_pil(__lowerCAmelCase ) if not return_dict: return (image,), "This is a local test" return ImagePipelineOutput(images=__lowerCAmelCase ), "This is a local test"
72
"""simple docstring""" from __future__ import annotations def snake_case_ ( A_ : str ): '''simple docstring''' return [ord(A_ ) - 96 for elem in plain] def snake_case_ ( A_ : list[int] ): '''simple docstring''' return "".join(chr(elem + 96 ) for elem in encoded ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : Dict = encode(input('''-> ''' ).strip().lower() ) print('''Encoded: ''', A_ ) print('''Decoded:''', decode(A_ ) ) if __name__ == "__main__": main()
72
1
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class __lowerCamelCase ( unittest.TestCase ): """simple docstring""" def a ( self : List[str] ) -> Optional[int]: lowerCAmelCase__ = "laion/clap-htsat-unfused" lowerCAmelCase__ = tempfile.mkdtemp() def a ( self : List[Any] , **SCREAMING_SNAKE_CASE__ : Dict ) -> Dict: return RobertaTokenizer.from_pretrained(self.checkpoint , **SCREAMING_SNAKE_CASE__ ) def a ( self : int , **SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[Any]: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **SCREAMING_SNAKE_CASE__ ) def a ( self : Any ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def a ( self : List[str] ) -> Union[str, Any]: lowerCAmelCase__ = self.get_tokenizer() lowerCAmelCase__ = self.get_feature_extractor() lowerCAmelCase__ = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) processor.save_pretrained(self.tmpdirname ) lowerCAmelCase__ = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , SCREAMING_SNAKE_CASE__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , SCREAMING_SNAKE_CASE__ ) def a ( self : Dict ) -> List[str]: lowerCAmelCase__ = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) lowerCAmelCase__ = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) lowerCAmelCase__ = self.get_feature_extractor(do_normalize=SCREAMING_SNAKE_CASE__ , padding_value=1.0 ) lowerCAmelCase__ = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=SCREAMING_SNAKE_CASE__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , SCREAMING_SNAKE_CASE__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , SCREAMING_SNAKE_CASE__ ) def a ( self : Any ) -> int: lowerCAmelCase__ = self.get_feature_extractor() lowerCAmelCase__ = self.get_tokenizer() lowerCAmelCase__ = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = floats_list((3, 1_000) ) lowerCAmelCase__ = feature_extractor(SCREAMING_SNAKE_CASE__ , return_tensors="np" ) lowerCAmelCase__ = processor(audios=SCREAMING_SNAKE_CASE__ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def a ( self : List[str] ) -> List[str]: lowerCAmelCase__ = self.get_feature_extractor() lowerCAmelCase__ = self.get_tokenizer() lowerCAmelCase__ = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = "This is a test string" lowerCAmelCase__ = processor(text=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = tokenizer(SCREAMING_SNAKE_CASE__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def a ( self : Any ) -> Optional[Any]: lowerCAmelCase__ = self.get_feature_extractor() lowerCAmelCase__ = self.get_tokenizer() lowerCAmelCase__ = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCAmelCase__ = processor.batch_decode(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = tokenizer.batch_decode(SCREAMING_SNAKE_CASE__ ) self.assertListEqual(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def a ( self : Union[str, Any] ) -> Tuple: lowerCAmelCase__ = self.get_feature_extractor() lowerCAmelCase__ = self.get_tokenizer() lowerCAmelCase__ = ClapProcessor(tokenizer=SCREAMING_SNAKE_CASE__ , feature_extractor=SCREAMING_SNAKE_CASE__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , )
362
import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig UpperCamelCase = logging.get_logger(__name__) # General docstring UpperCamelCase = 'PoolFormerConfig' # Base docstring UpperCamelCase = 'sail/poolformer_s12' UpperCamelCase = [1, 512, 7, 7] # Image classification docstring UpperCamelCase = 'sail/poolformer_s12' UpperCamelCase = 'tabby, tabby cat' UpperCamelCase = [ 'sail/poolformer_s12', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def _A ( lowerCAmelCase_ : Tuple , lowerCAmelCase_ : float = 0.0 , lowerCAmelCase_ : bool = False ): """simple docstring""" if drop_prob == 0.0 or not training: return input lowerCAmelCase__ = 1 - drop_prob lowerCAmelCase__ = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets lowerCAmelCase__ = keep_prob + torch.rand(lowerCAmelCase_ , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize lowerCAmelCase__ = input.div(lowerCAmelCase_ ) * random_tensor return output class __lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : Optional[float] = None ) -> None: super().__init__() lowerCAmelCase__ = drop_prob def a ( self : str , SCREAMING_SNAKE_CASE__ : torch.Tensor ) -> torch.Tensor: return drop_path(SCREAMING_SNAKE_CASE__ , self.drop_prob , self.training ) def a ( self : Optional[Any] ) -> str: return "p={}".format(self.drop_prob ) class __lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : str=None ) -> Optional[Any]: super().__init__() lowerCAmelCase__ = patch_size if isinstance(SCREAMING_SNAKE_CASE__ , collections.abc.Iterable ) else (patch_size, patch_size) lowerCAmelCase__ = stride if isinstance(SCREAMING_SNAKE_CASE__ , collections.abc.Iterable ) else (stride, stride) lowerCAmelCase__ = padding if isinstance(SCREAMING_SNAKE_CASE__ , collections.abc.Iterable ) else (padding, padding) lowerCAmelCase__ = nn.Convad(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , kernel_size=SCREAMING_SNAKE_CASE__ , stride=SCREAMING_SNAKE_CASE__ , padding=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = norm_layer(SCREAMING_SNAKE_CASE__ ) if norm_layer else nn.Identity() def a ( self : int , SCREAMING_SNAKE_CASE__ : List[Any] ) -> Union[str, Any]: lowerCAmelCase__ = self.projection(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.norm(SCREAMING_SNAKE_CASE__ ) return embeddings class __lowerCamelCase ( nn.GroupNorm ): """simple docstring""" def __init__( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[str] , **SCREAMING_SNAKE_CASE__ : Any ) -> Dict: super().__init__(1 , SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) class __lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : int ) -> List[Any]: super().__init__() lowerCAmelCase__ = nn.AvgPoolad(SCREAMING_SNAKE_CASE__ , stride=1 , padding=pool_size // 2 , count_include_pad=SCREAMING_SNAKE_CASE__ ) def a ( self : int , SCREAMING_SNAKE_CASE__ : List[Any] ) -> Optional[Any]: return self.pool(SCREAMING_SNAKE_CASE__ ) - hidden_states class __lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Tuple ) -> Dict: super().__init__() lowerCAmelCase__ = nn.Convad(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 ) lowerCAmelCase__ = nn.Convad(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 1 ) lowerCAmelCase__ = PoolFormerDropPath(SCREAMING_SNAKE_CASE__ ) if isinstance(config.hidden_act , SCREAMING_SNAKE_CASE__ ): lowerCAmelCase__ = ACTaFN[config.hidden_act] else: lowerCAmelCase__ = config.hidden_act def a ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> List[Any]: lowerCAmelCase__ = self.conva(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.act_fn(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.drop(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.conva(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.drop(SCREAMING_SNAKE_CASE__ ) return hidden_states class __lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : int ) -> Optional[Any]: super().__init__() lowerCAmelCase__ = PoolFormerPooling(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = PoolFormerOutput(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = PoolFormerGroupNorm(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = PoolFormerGroupNorm(SCREAMING_SNAKE_CASE__ ) # Useful for training neural nets lowerCAmelCase__ = PoolFormerDropPath(SCREAMING_SNAKE_CASE__ ) if drop_path > 0.0 else nn.Identity() lowerCAmelCase__ = config.use_layer_scale if config.use_layer_scale: lowerCAmelCase__ = nn.Parameter( config.layer_scale_init_value * torch.ones((SCREAMING_SNAKE_CASE__) ) , requires_grad=SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = nn.Parameter( config.layer_scale_init_value * torch.ones((SCREAMING_SNAKE_CASE__) ) , requires_grad=SCREAMING_SNAKE_CASE__ ) def a ( self : Dict , SCREAMING_SNAKE_CASE__ : str ) -> int: if self.use_layer_scale: lowerCAmelCase__ = self.pooling(self.before_norm(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection lowerCAmelCase__ = hidden_states + self.drop_path(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = () lowerCAmelCase__ = self.output(self.after_norm(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection lowerCAmelCase__ = hidden_states + self.drop_path(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = (output,) + outputs return outputs else: lowerCAmelCase__ = self.drop_path(self.pooling(self.before_norm(SCREAMING_SNAKE_CASE__ ) ) ) # First residual connection lowerCAmelCase__ = pooling_output + hidden_states lowerCAmelCase__ = () # Second residual connection inside the PoolFormerOutput block lowerCAmelCase__ = self.drop_path(self.output(self.after_norm(SCREAMING_SNAKE_CASE__ ) ) ) lowerCAmelCase__ = hidden_states + layer_output lowerCAmelCase__ = (output,) + outputs return outputs class __lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Dict ) -> Any: super().__init__() lowerCAmelCase__ = config # stochastic depth decay rule lowerCAmelCase__ = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings lowerCAmelCase__ = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) lowerCAmelCase__ = nn.ModuleList(SCREAMING_SNAKE_CASE__ ) # Transformer blocks lowerCAmelCase__ = [] lowerCAmelCase__ = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers lowerCAmelCase__ = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( SCREAMING_SNAKE_CASE__ , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = nn.ModuleList(SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : int=False , SCREAMING_SNAKE_CASE__ : List[str]=True ) -> Dict: lowerCAmelCase__ = () if output_hidden_states else None lowerCAmelCase__ = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): lowerCAmelCase__ , lowerCAmelCase__ = layers # Get patch embeddings from hidden_states lowerCAmelCase__ = embedding_layer(SCREAMING_SNAKE_CASE__ ) # Send the embeddings through the blocks for _, blk in enumerate(SCREAMING_SNAKE_CASE__ ): lowerCAmelCase__ = blk(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = layer_outputs[0] if output_hidden_states: lowerCAmelCase__ = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=SCREAMING_SNAKE_CASE__ , hidden_states=SCREAMING_SNAKE_CASE__ ) class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" snake_case__ = PoolFormerConfig snake_case__ = "poolformer" snake_case__ = "pixel_values" snake_case__ = True def a ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : int ) -> List[Any]: if isinstance(SCREAMING_SNAKE_CASE__ , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(SCREAMING_SNAKE_CASE__ , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def a ( self : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Union[str, Any]=False ) -> Tuple: if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): lowerCAmelCase__ = value UpperCamelCase = R'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' UpperCamelCase = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`PoolFormerImageProcessor.__call__`] for details.\n' @add_start_docstrings( "The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top." , UpperCamelCase__ , ) class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : int ) -> Optional[Any]: super().__init__(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = config lowerCAmelCase__ = PoolFormerEncoder(SCREAMING_SNAKE_CASE__ ) # Initialize weights and apply final processing self.post_init() def a ( self : Optional[int] ) -> Optional[Any]: return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(SCREAMING_SNAKE_CASE__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=SCREAMING_SNAKE_CASE__ , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def a ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[torch.FloatTensor] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , ) -> Union[Tuple, BaseModelOutputWithNoAttention]: lowerCAmelCase__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) lowerCAmelCase__ = self.encoder( SCREAMING_SNAKE_CASE__ , output_hidden_states=SCREAMING_SNAKE_CASE__ , return_dict=SCREAMING_SNAKE_CASE__ , ) lowerCAmelCase__ = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=SCREAMING_SNAKE_CASE__ , hidden_states=encoder_outputs.hidden_states , ) class __lowerCamelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : List[Any] ) -> List[str]: super().__init__() lowerCAmelCase__ = nn.Linear(config.hidden_size , config.hidden_size ) def a ( self : Any , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> int: lowerCAmelCase__ = self.dense(SCREAMING_SNAKE_CASE__ ) return output @add_start_docstrings( "\n PoolFormer Model transformer with an image classification head on top\n " , UpperCamelCase__ , ) class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : List[Any] ) -> Optional[int]: super().__init__(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = config.num_labels lowerCAmelCase__ = PoolFormerModel(SCREAMING_SNAKE_CASE__ ) # Final norm lowerCAmelCase__ = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head lowerCAmelCase__ = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(SCREAMING_SNAKE_CASE__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=SCREAMING_SNAKE_CASE__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def a ( self : Tuple , SCREAMING_SNAKE_CASE__ : Optional[torch.FloatTensor] = None , SCREAMING_SNAKE_CASE__ : Optional[torch.LongTensor] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , SCREAMING_SNAKE_CASE__ : Optional[bool] = None , ) -> Union[Tuple, ImageClassifierOutputWithNoAttention]: lowerCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase__ = self.poolformer( SCREAMING_SNAKE_CASE__ , output_hidden_states=SCREAMING_SNAKE_CASE__ , return_dict=SCREAMING_SNAKE_CASE__ , ) lowerCAmelCase__ = outputs[0] lowerCAmelCase__ = self.classifier(self.norm(SCREAMING_SNAKE_CASE__ ).mean([-2, -1] ) ) lowerCAmelCase__ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowerCAmelCase__ = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowerCAmelCase__ = "single_label_classification" else: lowerCAmelCase__ = "multi_label_classification" if self.config.problem_type == "regression": lowerCAmelCase__ = MSELoss() if self.num_labels == 1: lowerCAmelCase__ = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowerCAmelCase__ = loss_fct(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) elif self.config.problem_type == "single_label_classification": lowerCAmelCase__ = CrossEntropyLoss() lowerCAmelCase__ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowerCAmelCase__ = BCEWithLogitsLoss() lowerCAmelCase__ = loss_fct(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if not return_dict: lowerCAmelCase__ = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=SCREAMING_SNAKE_CASE__ , logits=SCREAMING_SNAKE_CASE__ , hidden_states=outputs.hidden_states )
221
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class __magic_name__ ( metaclass=snake_case__ ): '''simple docstring''' __UpperCamelCase = ["torch", "scipy"] def __init__( self , *_a , **_a ): """simple docstring""" requires_backends(self , ["""torch""", """scipy"""] ) @classmethod def _lowerCAmelCase ( cls , *_a , **_a ): """simple docstring""" requires_backends(cls , ["""torch""", """scipy"""] ) @classmethod def _lowerCAmelCase ( cls , *_a , **_a ): """simple docstring""" requires_backends(cls , ["""torch""", """scipy"""] )
291
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : List[Any] = { """facebook/encodec_24khz""": """https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json""", """facebook/encodec_48khz""": """https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json""", } class lowerCamelCase_ (snake_case__ ): '''simple docstring''' __UpperCamelCase: str = "encodec" def __init__( self : Optional[int] , A : Union[str, Any]=[1.5, 3.0, 6.0, 12.0, 24.0] , A : List[Any]=24000 , A : Union[str, Any]=1 , A : List[Any]=False , A : Optional[int]=None , A : int=None , A : str=128 , A : List[Any]=32 , A : List[Any]=1 , A : int=[8, 5, 4, 2] , A : Optional[int]="weight_norm" , A : List[Any]=7 , A : Any=7 , A : Dict=3 , A : Optional[int]=2 , A : Dict=True , A : Dict="reflect" , A : Any=2 , A : Dict=2 , A : str=1.0 , A : Optional[int]=1024 , A : Any=None , A : Any=True , **A : str , ): _UpperCAmelCase : Optional[int] = target_bandwidths _UpperCAmelCase : List[str] = sampling_rate _UpperCAmelCase : Optional[int] = audio_channels _UpperCAmelCase : str = normalize _UpperCAmelCase : int = chunk_length_s _UpperCAmelCase : str = overlap _UpperCAmelCase : Optional[Any] = hidden_size _UpperCAmelCase : int = num_filters _UpperCAmelCase : Optional[Any] = num_residual_layers _UpperCAmelCase : Optional[int] = upsampling_ratios _UpperCAmelCase : int = norm_type _UpperCAmelCase : List[Any] = kernel_size _UpperCAmelCase : List[Any] = last_kernel_size _UpperCAmelCase : List[Any] = residual_kernel_size _UpperCAmelCase : List[str] = dilation_growth_rate _UpperCAmelCase : Dict = use_causal_conv _UpperCAmelCase : Tuple = pad_mode _UpperCAmelCase : Tuple = compress _UpperCAmelCase : List[str] = num_lstm_layers _UpperCAmelCase : List[Any] = trim_right_ratio _UpperCAmelCase : int = codebook_size _UpperCAmelCase : Optional[Any] = codebook_dim if codebook_dim is not None else hidden_size _UpperCAmelCase : Optional[int] = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( F"""self.norm_type must be one of `\"weight_norm\"`, `\"time_group_norm\"`), got {self.norm_type}""" ) super().__init__(**A ) @property def _A ( self : Any ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def _A ( self : Union[str, Any] ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def _A ( self : Union[str, Any] ): _UpperCAmelCase : Dict = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def _A ( self : str ): return int(1000 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
31
0
"""simple docstring""" import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig lowerCamelCase = { """facebook/maskformer-swin-base-ade""": ( """https://huggingface.co/facebook/maskformer-swin-base-ade/blob/main/config.json""" ) # See all MaskFormer models at https://huggingface.co/models?filter=maskformer } lowerCamelCase = logging.get_logger(__name__) class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = '''maskformer''' UpperCamelCase = {'''hidden_size''': '''mask_feature_size'''} UpperCamelCase = ['''resnet''', '''swin'''] UpperCamelCase = ['''detr'''] def __init__( self : Optional[int] , _UpperCAmelCase : int = 256 , _UpperCAmelCase : int = 256 , _UpperCAmelCase : float = 0.1 , _UpperCAmelCase : bool = False , _UpperCAmelCase : Optional[Dict] = None , _UpperCAmelCase : Optional[Dict] = None , _UpperCAmelCase : float = 0.02 , _UpperCAmelCase : float = 1.0 , _UpperCAmelCase : float = 1.0 , _UpperCAmelCase : float = 1.0 , _UpperCAmelCase : float = 20.0 , _UpperCAmelCase : Optional[bool] = None , **_UpperCAmelCase : Dict , ) -> Union[str, Any]: '''simple docstring''' if backbone_config is None: # fall back to https://huggingface.co/microsoft/swin-base-patch4-window12-384-in22k UpperCAmelCase_ = SwinConfig( image_size=384 , in_channels=3 , patch_size=4 , embed_dim=128 , depths=[2, 2, 18, 2] , num_heads=[4, 8, 16, 32] , window_size=12 , drop_path_rate=0.3 , out_features=["stage1", "stage2", "stage3", "stage4"] , ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): UpperCAmelCase_ = backbone_config.pop("model_type" ) UpperCAmelCase_ = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase_ = config_class.from_dict(_UpperCAmelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with MaskFormer. """ F"""Supported model types: {','.join(self.backbones_supported )}""" ) if decoder_config is None: # fall back to https://huggingface.co/facebook/detr-resnet-50 UpperCAmelCase_ = DetrConfig() else: # verify that the decoder is supported UpperCAmelCase_ = ( decoder_config.pop("model_type" ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else decoder_config.model_type ) if decoder_type not in self.decoders_supported: raise ValueError( F"""Transformer Decoder {decoder_type} not supported, please use one of""" F""" {','.join(self.decoders_supported )}""" ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): UpperCAmelCase_ = CONFIG_MAPPING[decoder_type] UpperCAmelCase_ = config_class.from_dict(_UpperCAmelCase ) UpperCAmelCase_ = backbone_config UpperCAmelCase_ = decoder_config # main feature dimension for the model UpperCAmelCase_ = fpn_feature_size UpperCAmelCase_ = mask_feature_size # initializer UpperCAmelCase_ = init_std UpperCAmelCase_ = init_xavier_std # Hungarian matcher && loss UpperCAmelCase_ = cross_entropy_weight UpperCAmelCase_ = dice_weight UpperCAmelCase_ = mask_weight UpperCAmelCase_ = use_auxiliary_loss UpperCAmelCase_ = no_object_weight UpperCAmelCase_ = output_auxiliary_logits UpperCAmelCase_ = self.decoder_config.encoder_attention_heads UpperCAmelCase_ = self.decoder_config.num_hidden_layers super().__init__(**_UpperCAmelCase ) @classmethod def lowercase__ ( cls : Any , _UpperCAmelCase : PretrainedConfig , _UpperCAmelCase : PretrainedConfig , **_UpperCAmelCase : List[Any] ) -> Optional[int]: '''simple docstring''' return cls( backbone_config=_UpperCAmelCase , decoder_config=_UpperCAmelCase , **_UpperCAmelCase , ) def lowercase__ ( self : List[str] ) -> Dict[str, any]: '''simple docstring''' UpperCAmelCase_ = copy.deepcopy(self.__dict__ ) UpperCAmelCase_ = self.backbone_config.to_dict() UpperCAmelCase_ = self.decoder_config.to_dict() UpperCAmelCase_ = self.__class__.model_type return output
241
"""simple docstring""" from maths.prime_check import is_prime def a__ ( lowerCAmelCase__ ): if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowerCAmelCase__ ) if is_prime(lowerCAmelCase__ ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
241
1
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) __snake_case = '''hf-internal-testing/tiny-random-bert''' __snake_case = os.path.join(TRANSFORMERS_CACHE, '''models--hf-internal-testing--tiny-random-bert''') __snake_case = '''9b8c223d42b2188cb49d29af482996f9d0f3e5a6''' class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : int =cached_file(UpperCamelCase_ , UpperCamelCase_ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCamelCase_ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCamelCase_ , UpperCamelCase_ ) ) ) with open(os.path.join(UpperCamelCase_ , '''refs''' , '''main''' ) ) as f: UpperCAmelCase : Union[str, Any] =f.read() self.assertEqual(UpperCamelCase_ , os.path.join(UpperCamelCase_ , '''snapshots''' , UpperCamelCase_ , UpperCamelCase_ ) ) self.assertTrue(os.path.isfile(UpperCamelCase_ ) ) # File is cached at the same place the second time. UpperCAmelCase : Optional[int] =cached_file(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) # Using a specific revision to test the full commit hash. UpperCAmelCase : Union[str, Any] =cached_file(UpperCamelCase_ , UpperCamelCase_ , revision='''9b8c223''' ) self.assertEqual(UpperCamelCase_ , os.path.join(UpperCamelCase_ , '''snapshots''' , UpperCamelCase_ , UpperCamelCase_ ) ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' with self.assertRaisesRegex(UpperCamelCase_ , '''is not a valid model identifier''' ): UpperCAmelCase : List[str] =cached_file('''tiny-random-bert''' , UpperCamelCase_ ) with self.assertRaisesRegex(UpperCamelCase_ , '''is not a valid git identifier''' ): UpperCAmelCase : Any =cached_file(UpperCamelCase_ , UpperCamelCase_ , revision='''aaaa''' ) with self.assertRaisesRegex(UpperCamelCase_ , '''does not appear to have a file named''' ): UpperCAmelCase : Dict =cached_file(UpperCamelCase_ , '''conf''' ) def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' with self.assertRaisesRegex(UpperCamelCase_ , '''does not appear to have a file named''' ): UpperCAmelCase : Optional[int] =cached_file(UpperCamelCase_ , '''conf''' ) with open(os.path.join(UpperCamelCase_ , '''refs''' , '''main''' ) ) as f: UpperCAmelCase : Any =f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCamelCase_ , '''.no_exist''' , UpperCamelCase_ , '''conf''' ) ) ) UpperCAmelCase : Tuple =cached_file(UpperCamelCase_ , '''conf''' , _raise_exceptions_for_missing_entries=UpperCamelCase_ ) self.assertIsNone(UpperCamelCase_ ) UpperCAmelCase : List[Any] =cached_file(UpperCamelCase_ , '''conf''' , local_files_only=UpperCamelCase_ , _raise_exceptions_for_missing_entries=UpperCamelCase_ ) self.assertIsNone(UpperCamelCase_ ) UpperCAmelCase : Union[str, Any] =mock.Mock() UpperCAmelCase : List[str] =500 UpperCAmelCase : Optional[int] ={} UpperCAmelCase : Any =HTTPError UpperCAmelCase : Tuple ={} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=UpperCamelCase_ ) as mock_head: UpperCAmelCase : Optional[Any] =cached_file(UpperCamelCase_ , '''conf''' , _raise_exceptions_for_connection_errors=UpperCamelCase_ ) self.assertIsNone(UpperCamelCase_ ) # This check we did call the fake head request mock_head.assert_called() def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCamelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCamelCase_ ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , UpperCamelCase_ ) ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCamelCase_ , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , UpperCamelCase_ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCamelCase_ , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , UpperCamelCase_ , revision='''ahaha''' ) UpperCAmelCase : str =get_file_from_repo('''bert-base-cased''' , UpperCamelCase_ ) # The name is the cached name which is not very easy to test, so instead we load the content. UpperCAmelCase : Dict =json.loads(open(UpperCamelCase_ , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 768 ) def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase : List[Any] =Path(UpperCamelCase_ ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(UpperCamelCase_ , '''a.txt''' ) , str(UpperCamelCase_ ) ) self.assertIsNone(get_file_from_repo(UpperCamelCase_ , '''b.txt''' ) )
348
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class UpperCAmelCase_ ( unittest.TestCase ): def _lowerCamelCase ( self ) -> Union[str, Any]: __lowercase : Dict = tempfile.mkdtemp() __lowercase : Any = BlipImageProcessor() __lowercase : Optional[int] = GPTaTokenizer.from_pretrained('''hf-internal-testing/tiny-random-GPT2Model''' ) __lowercase : str = BertTokenizerFast.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) __lowercase : str = InstructBlipProcessor(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) processor.save_pretrained(self.tmpdirname ) def _lowerCamelCase ( self , **UpperCamelCase_ ) -> Any: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase_ ).tokenizer def _lowerCamelCase ( self , **UpperCamelCase_ ) -> List[str]: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase_ ).image_processor def _lowerCamelCase ( self , **UpperCamelCase_ ) -> List[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase_ ).qformer_tokenizer def _lowerCamelCase ( self ) -> Tuple: shutil.rmtree(self.tmpdirname ) def _lowerCamelCase ( self ) -> Any: __lowercase : Optional[int] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta )] __lowercase : Any = [Image.fromarray(np.moveaxis(UpperCamelCase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowerCamelCase ( self ) -> str: __lowercase : Any = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) __lowercase : List[str] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __lowercase : Dict = self.get_image_processor(do_normalize=UpperCamelCase_ , padding_value=1.0 ) __lowercase : int = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=UpperCamelCase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCamelCase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCamelCase_ ) self.assertIsInstance(processor.qformer_tokenizer , UpperCamelCase_ ) def _lowerCamelCase ( self ) -> Any: __lowercase : Any = self.get_image_processor() __lowercase : str = self.get_tokenizer() __lowercase : Any = self.get_qformer_tokenizer() __lowercase : List[str] = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowercase : int = self.prepare_image_inputs() __lowercase : Union[str, Any] = image_processor(UpperCamelCase_ , return_tensors='''np''' ) __lowercase : Tuple = processor(images=UpperCamelCase_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowerCamelCase ( self ) -> str: __lowercase : str = self.get_image_processor() __lowercase : Dict = self.get_tokenizer() __lowercase : Optional[Any] = self.get_qformer_tokenizer() __lowercase : List[str] = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowercase : Dict = '''lower newer''' __lowercase : int = processor(text=UpperCamelCase_ ) __lowercase : List[str] = tokenizer(UpperCamelCase_ , return_token_type_ids=UpperCamelCase_ ) __lowercase : Union[str, Any] = qformer_tokenizer(UpperCamelCase_ , return_token_type_ids=UpperCamelCase_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['''qformer_''' + key] ) def _lowerCamelCase ( self ) -> List[str]: __lowercase : Union[str, Any] = self.get_image_processor() __lowercase : Union[str, Any] = self.get_tokenizer() __lowercase : Optional[int] = self.get_qformer_tokenizer() __lowercase : List[str] = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowercase : Optional[int] = '''lower newer''' __lowercase : Any = self.prepare_image_inputs() __lowercase : List[Any] = processor(text=UpperCamelCase_ , images=UpperCamelCase_ ) self.assertListEqual( list(inputs.keys() ) , ['''input_ids''', '''attention_mask''', '''qformer_input_ids''', '''qformer_attention_mask''', '''pixel_values'''] , ) # test if it raises when no input is passed with pytest.raises(UpperCamelCase_ ): processor() def _lowerCamelCase ( self ) -> Dict: __lowercase : Any = self.get_image_processor() __lowercase : List[str] = self.get_tokenizer() __lowercase : Any = self.get_qformer_tokenizer() __lowercase : Tuple = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowercase : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowercase : List[str] = processor.batch_decode(UpperCamelCase_ ) __lowercase : Union[str, Any] = tokenizer.batch_decode(UpperCamelCase_ ) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_ ) def _lowerCamelCase ( self ) -> List[str]: __lowercase : List[str] = self.get_image_processor() __lowercase : List[str] = self.get_tokenizer() __lowercase : List[Any] = self.get_qformer_tokenizer() __lowercase : Optional[Any] = InstructBlipProcessor( tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_ , qformer_tokenizer=UpperCamelCase_ ) __lowercase : Any = '''lower newer''' __lowercase : Union[str, Any] = self.prepare_image_inputs() __lowercase : Union[str, Any] = processor(text=UpperCamelCase_ , images=UpperCamelCase_ ) self.assertListEqual( list(inputs.keys() ) , ['''input_ids''', '''attention_mask''', '''qformer_input_ids''', '''qformer_attention_mask''', '''pixel_values'''] , )
249
0
import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class __A : def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=13 , UpperCAmelCase_=7 , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=True , UpperCAmelCase_=99 , UpperCAmelCase_=32 , UpperCAmelCase_=5 , UpperCAmelCase_=4 , UpperCAmelCase_=37 , UpperCAmelCase_="gelu" , UpperCAmelCase_=0.1 , UpperCAmelCase_=0.1 , UpperCAmelCase_=512 , UpperCAmelCase_=16 , UpperCAmelCase_=2 , UpperCAmelCase_=0.0_2 , UpperCAmelCase_=3 , UpperCAmelCase_=4 , UpperCAmelCase_=None , ): lowerCamelCase =parent lowerCamelCase =batch_size lowerCamelCase =seq_length lowerCamelCase =is_training lowerCamelCase =use_input_mask lowerCamelCase =use_token_type_ids lowerCamelCase =use_labels lowerCamelCase =vocab_size lowerCamelCase =hidden_size lowerCamelCase =num_hidden_layers lowerCamelCase =num_attention_heads lowerCamelCase =intermediate_size lowerCamelCase =hidden_act lowerCamelCase =hidden_dropout_prob lowerCamelCase =attention_probs_dropout_prob lowerCamelCase =max_position_embeddings lowerCamelCase =type_vocab_size lowerCamelCase =type_sequence_label_size lowerCamelCase =initializer_range lowerCamelCase =num_labels lowerCamelCase =num_choices lowerCamelCase =scope def _snake_case ( self ): lowerCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase =None if self.use_input_mask: lowerCamelCase =random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase =None if self.use_token_type_ids: lowerCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase =None lowerCamelCase =None lowerCamelCase =None if self.use_labels: lowerCamelCase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase =ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase =self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _snake_case ( self ): return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCAmelCase_ , initializer_range=self.initializer_range , ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =NystromformerModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCamelCase =model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ ) lowerCamelCase =model(UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ ) lowerCamelCase =model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =NystromformerForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCamelCase =model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =NystromformerForQuestionAnswering(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCamelCase =model( UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , start_positions=UpperCAmelCase_ , end_positions=UpperCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =self.num_labels lowerCamelCase =NystromformerForSequenceClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCamelCase =model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =self.num_labels lowerCamelCase =NystromformerForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCamelCase =model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _snake_case ( self , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase =self.num_choices lowerCamelCase =NystromformerForMultipleChoice(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() lowerCamelCase =input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCamelCase =token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCamelCase =input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCamelCase =model( UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _snake_case ( self ): lowerCamelCase =self.prepare_config_and_inputs() ( ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ( lowerCamelCase ) , ) =config_and_inputs lowerCamelCase ={"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __A ( a , a , unittest.TestCase ): __A = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) __A = ( { """feature-extraction""": NystromformerModel, """fill-mask""": NystromformerForMaskedLM, """question-answering""": NystromformerForQuestionAnswering, """text-classification""": NystromformerForSequenceClassification, """token-classification""": NystromformerForTokenClassification, """zero-shot""": NystromformerForSequenceClassification, } if is_torch_available() else {} ) __A = False __A = False def _snake_case ( self ): lowerCamelCase =NystromformerModelTester(self ) lowerCamelCase =ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def _snake_case ( self ): self.config_tester.run_common_tests() def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCamelCase =type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*UpperCAmelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def _snake_case ( self ): lowerCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def _snake_case ( self ): for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase =NystromformerModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) @require_torch class __A ( unittest.TestCase ): @slow def _snake_case ( self ): lowerCamelCase =NystromformerModel.from_pretrained("""uw-madison/nystromformer-512""" ) lowerCamelCase =torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): lowerCamelCase =model(UpperCAmelCase_ )[0] lowerCamelCase =torch.Size((1, 6, 768) ) self.assertEqual(output.shape , UpperCAmelCase_ ) lowerCamelCase =torch.tensor( [[[-0.4_5_3_2, -0.0_9_3_6, 0.5_1_3_7], [-0.2_6_7_6, 0.0_6_2_8, 0.6_1_8_6], [-0.3_6_2_9, -0.1_7_2_6, 0.4_7_1_6]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1E-4 ) ) @slow def _snake_case ( self ): lowerCamelCase ="""the [MASK] of Belgium is Brussels""" lowerCamelCase =AutoTokenizer.from_pretrained("""uw-madison/nystromformer-512""" ) lowerCamelCase =NystromformerForMaskedLM.from_pretrained("""uw-madison/nystromformer-512""" ) lowerCamelCase =tokenizer(UpperCAmelCase_ , return_tensors="""pt""" ) with torch.no_grad(): lowerCamelCase =model(encoding.input_ids ).logits lowerCamelCase =token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(UpperCAmelCase_ ) , """capital""" )
262
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process UpperCAmelCase__ : Union[str, Any] =logging.getLogger(__name__) def _lowercase ( _UpperCAmelCase , _UpperCAmelCase ) -> int: return (preds == labels).mean() @dataclass class __A : __A = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __A = field( default=a , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __A = field( default=a , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __A = field( default=a , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class __A : __A = field(metadata={"""help""": """The name of the task to train on: """ + """, """.join(processors.keys() )} ) __A = field(metadata={"""help""": """Should contain the data files for the task."""} ) __A = field( default=1_28 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __A = field( default=a , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def _lowercase ( ) -> Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCamelCase =HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowerCamelCase , lowerCamelCase , lowerCamelCase =parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , _UpperCAmelCase ) # Set seed set_seed(training_args.seed ) try: lowerCamelCase =processors[data_args.task_name]() lowerCamelCase =processor.get_labels() lowerCamelCase =len(_UpperCAmelCase ) except KeyError: raise ValueError("""Task not found: %s""" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCamelCase =AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=_UpperCAmelCase , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) lowerCamelCase =AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) lowerCamelCase =AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_UpperCAmelCase , cache_dir=model_args.cache_dir , ) # Get datasets lowerCamelCase =( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_UpperCAmelCase , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) lowerCamelCase =( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=_UpperCAmelCase , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(_UpperCAmelCase ) -> Dict: lowerCamelCase =np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(_UpperCAmelCase , p.label_ids )} # Data collator lowerCamelCase =DataCollatorWithPadding(_UpperCAmelCase , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer lowerCamelCase =Trainer( model=_UpperCAmelCase , args=_UpperCAmelCase , train_dataset=_UpperCAmelCase , eval_dataset=_UpperCAmelCase , compute_metrics=_UpperCAmelCase , data_collator=_UpperCAmelCase , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowerCamelCase ={} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) lowerCamelCase =trainer.evaluate() lowerCamelCase =os.path.join(training_args.output_dir , """eval_results.txt""" ) if trainer.is_world_master(): with open(_UpperCAmelCase , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key, value in result.items(): logger.info(""" %s = %s""" , _UpperCAmelCase , _UpperCAmelCase ) writer.write("""%s = %s\n""" % (key, value) ) results.update(_UpperCAmelCase ) return results def _lowercase ( _UpperCAmelCase ) -> Union[str, Any]: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
262
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : str = logging.get_logger(__name__) _A : List[Any] = { '''uclanlp/visualbert-vqa''': '''https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json''', '''uclanlp/visualbert-vqa-pre''': '''https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json''', '''uclanlp/visualbert-vqa-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json''' ), '''uclanlp/visualbert-vcr''': '''https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json''', '''uclanlp/visualbert-vcr-pre''': '''https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json''', '''uclanlp/visualbert-vcr-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json''' ), '''uclanlp/visualbert-nlvr2''': '''https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json''', '''uclanlp/visualbert-nlvr2-pre''': '''https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json''', '''uclanlp/visualbert-nlvr2-coco-pre''': ( '''https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json''' ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class _lowercase ( UpperCAmelCase__ ): '''simple docstring''' _SCREAMING_SNAKE_CASE : Union[str, Any] = """visual_bert""" def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : int=3_05_22 , SCREAMING_SNAKE_CASE__ : str=7_68 , SCREAMING_SNAKE_CASE__ : Optional[int]=5_12 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=12 , SCREAMING_SNAKE_CASE__ : Any=12 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=30_72 , SCREAMING_SNAKE_CASE__ : int="gelu" , SCREAMING_SNAKE_CASE__ : List[Any]=0.1 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0.1 , SCREAMING_SNAKE_CASE__ : Any=5_12 , SCREAMING_SNAKE_CASE__ : Tuple=2 , SCREAMING_SNAKE_CASE__ : Optional[Any]=0.0_2 , SCREAMING_SNAKE_CASE__ : int=1e-1_2 , SCREAMING_SNAKE_CASE__ : Optional[int]=False , SCREAMING_SNAKE_CASE__ : Any=True , SCREAMING_SNAKE_CASE__ : int=1 , SCREAMING_SNAKE_CASE__ : Any=0 , SCREAMING_SNAKE_CASE__ : str=2 , **SCREAMING_SNAKE_CASE__ : List[Any] , ) -> str: super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ , bos_token_id=SCREAMING_SNAKE_CASE__ , eos_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = hidden_size __lowerCAmelCase = visual_embedding_dim __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = type_vocab_size __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = bypass_transformer __lowerCAmelCase = special_visual_initialize
229
'''simple docstring''' import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler _A : Optional[Any] = 16 _A : Union[str, Any] = 32 def UpperCamelCase_ ( snake_case_ : List[str] ) -> str: '''simple docstring''' return int(x / 2**20 ) class _lowercase : '''simple docstring''' def __enter__( self : List[Any] ) -> int: gc.collect() torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() # reset the peak gauge to zero __lowerCAmelCase = torch.cuda.memory_allocated() return self def __exit__( self : Tuple , *SCREAMING_SNAKE_CASE__ : List[Any] ) -> Optional[Any]: gc.collect() torch.cuda.empty_cache() __lowerCAmelCase = torch.cuda.memory_allocated() __lowerCAmelCase = torch.cuda.max_memory_allocated() __lowerCAmelCase = bamb(self.end - self.begin ) __lowerCAmelCase = bamb(self.peak - self.begin ) # print(f"delta used/peak {self.used:4d}/{self.peaked:4d}") def UpperCamelCase_ ( snake_case_ : Accelerator , snake_case_ : int = 16 , snake_case_ : str = "bert-base-cased" , snake_case_ : int = 3_20 , snake_case_ : int = 1_60 , ) -> Optional[int]: '''simple docstring''' __lowerCAmelCase = AutoTokenizer.from_pretrained(snake_case_ ) __lowerCAmelCase = load_dataset( """glue""" , """mrpc""" , split={"""train""": f"""train[:{n_train}]""", """validation""": f"""validation[:{n_val}]"""} ) def tokenize_function(snake_case_ : List[Any] ): # max_length=None => use the model max length (it's actually the default) __lowerCAmelCase = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case_ , max_length=snake_case_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __lowerCAmelCase = datasets.map( snake_case_ , batched=snake_case_ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , load_from_cache_file=snake_case_ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __lowerCAmelCase = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(snake_case_ : List[Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(snake_case_ , padding="""max_length""" , max_length=1_28 , return_tensors="""pt""" ) return tokenizer.pad(snake_case_ , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. __lowerCAmelCase = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=snake_case_ ) __lowerCAmelCase = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case_ , collate_fn=snake_case_ , batch_size=snake_case_ ) return train_dataloader, eval_dataloader def UpperCamelCase_ ( snake_case_ : List[Any] , snake_case_ : Tuple ) -> Optional[int]: '''simple docstring''' __lowerCAmelCase = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __lowerCAmelCase = config["""lr"""] __lowerCAmelCase = int(config["""num_epochs"""] ) __lowerCAmelCase = int(config["""seed"""] ) __lowerCAmelCase = int(config["""batch_size"""] ) __lowerCAmelCase = args.model_name_or_path set_seed(snake_case_ ) __lowerCAmelCase , __lowerCAmelCase = get_dataloaders(snake_case_ , snake_case_ , snake_case_ , args.n_train , args.n_val ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __lowerCAmelCase = AutoModelForSequenceClassification.from_pretrained(snake_case_ , return_dict=snake_case_ ) # Instantiate optimizer __lowerCAmelCase = ( AdamW if accelerator.state.deepspeed_plugin is None or """optimizer""" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __lowerCAmelCase = optimizer_cls(params=model.parameters() , lr=snake_case_ ) if accelerator.state.deepspeed_plugin is not None: __lowerCAmelCase = accelerator.state.deepspeed_plugin.deepspeed_config[ """gradient_accumulation_steps""" ] else: __lowerCAmelCase = 1 __lowerCAmelCase = (len(snake_case_ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __lowerCAmelCase = get_linear_schedule_with_warmup( optimizer=snake_case_ , num_warmup_steps=0 , num_training_steps=snake_case_ , ) else: __lowerCAmelCase = DummyScheduler(snake_case_ , total_num_steps=snake_case_ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = accelerator.prepare( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) # We need to keep track of how many total steps we have iterated over __lowerCAmelCase = 0 # We also need to keep track of the stating epoch so files are named properly __lowerCAmelCase = 0 # Now we train the model __lowerCAmelCase = {} for epoch in range(snake_case_ , snake_case_ ): with TorchTracemalloc() as tracemalloc: model.train() for step, batch in enumerate(snake_case_ ): __lowerCAmelCase = model(**snake_case_ ) __lowerCAmelCase = outputs.loss __lowerCAmelCase = loss / gradient_accumulation_steps accelerator.backward(snake_case_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 # Printing the GPU memory usage details such as allocated memory, peak memory, and total memory usage accelerator.print("""Memory before entering the train : {}""".format(bamb(tracemalloc.begin ) ) ) accelerator.print("""Memory consumed at the end of the train (end-begin): {}""".format(tracemalloc.used ) ) accelerator.print("""Peak Memory consumed during the train (max-begin): {}""".format(tracemalloc.peaked ) ) accelerator.print( """Total Peak Memory consumed during the train (max): {}""".format( tracemalloc.peaked + bamb(tracemalloc.begin ) ) ) __lowerCAmelCase = tracemalloc.peaked + bamb(tracemalloc.begin ) if args.peak_memory_upper_bound is not None: assert ( train_total_peak_memory[f"""epoch-{epoch}"""] <= args.peak_memory_upper_bound ), "Peak memory usage exceeded the upper bound" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , """peak_memory_utilization.json""" ) , """w""" ) as f: json.dump(snake_case_ , snake_case_ ) def UpperCamelCase_ ( ) -> Any: '''simple docstring''' __lowerCAmelCase = argparse.ArgumentParser(description="""Simple example of training script tracking peak GPU memory usage.""" ) parser.add_argument( """--model_name_or_path""" , type=snake_case_ , default="""bert-base-cased""" , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=snake_case_ , ) parser.add_argument( """--output_dir""" , type=snake_case_ , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--peak_memory_upper_bound""" , type=snake_case_ , default=snake_case_ , help="""The upper bound of peak memory usage in MB. If set, the training will throw an error if the peak memory usage exceeds this value.""" , ) parser.add_argument( """--n_train""" , type=snake_case_ , default=3_20 , help="""Number of training examples to use.""" , ) parser.add_argument( """--n_val""" , type=snake_case_ , default=1_60 , help="""Number of validation examples to use.""" , ) parser.add_argument( """--num_epochs""" , type=snake_case_ , default=1 , help="""Number of train epochs.""" , ) __lowerCAmelCase = parser.parse_args() __lowerCAmelCase = {"""lr""": 2E-5, """num_epochs""": args.num_epochs, """seed""": 42, """batch_size""": 16} training_function(snake_case_ , snake_case_ ) if __name__ == "__main__": main()
229
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to properly calculate the metrics on the # validation dataset when in a distributed system, and builds off the # `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __a = 16 __a = 32 def lowerCamelCase__ ( _lowercase , _lowercase = 16 ): '''simple docstring''' UpperCAmelCase_ = AutoTokenizer.from_pretrained('''bert-base-cased''' ) UpperCAmelCase_ = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(_lowercase ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase_ = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__lowerCAmelCase , max_length=__lowerCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): UpperCAmelCase_ = datasets.map( __lowerCAmelCase , batched=__lowerCAmelCase , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase_ = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(_lowercase ): # On TPU it's best to pad everything to the same length or training will be very slow. UpperCAmelCase_ = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": UpperCAmelCase_ = 16 elif accelerator.mixed_precision != "no": UpperCAmelCase_ = 8 else: UpperCAmelCase_ = None return tokenizer.pad( __lowerCAmelCase , padding='''longest''' , max_length=__lowerCAmelCase , pad_to_multiple_of=__lowerCAmelCase , return_tensors='''pt''' , ) # Instantiate dataloaders. UpperCAmelCase_ = DataLoader( tokenized_datasets['''train'''] , shuffle=__lowerCAmelCase , collate_fn=__lowerCAmelCase , batch_size=__lowerCAmelCase ) UpperCAmelCase_ = DataLoader( tokenized_datasets['''validation'''] , shuffle=__lowerCAmelCase , collate_fn=__lowerCAmelCase , batch_size=__lowerCAmelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders __a = mocked_dataloaders # noqa: F811 def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __lowerCAmelCase ) == "1": UpperCAmelCase_ = 2 # Initialize accelerator UpperCAmelCase_ = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase_ = config['''lr'''] UpperCAmelCase_ = int(config['''num_epochs'''] ) UpperCAmelCase_ = int(config['''seed'''] ) UpperCAmelCase_ = int(config['''batch_size'''] ) UpperCAmelCase_ = evaluate.load('''glue''' , '''mrpc''' ) # If the batch size is too big we use gradient accumulation UpperCAmelCase_ = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: UpperCAmelCase_ = batch_size // MAX_GPU_BATCH_SIZE UpperCAmelCase_ = MAX_GPU_BATCH_SIZE set_seed(__lowerCAmelCase ) UpperCAmelCase_ = get_dataloaders(__lowerCAmelCase , __lowerCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase_ = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__lowerCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCAmelCase_ = model.to(accelerator.device ) # Instantiate optimizer UpperCAmelCase_ = AdamW(params=model.parameters() , lr=__lowerCAmelCase ) # Instantiate scheduler UpperCAmelCase_ = get_linear_schedule_with_warmup( optimizer=__lowerCAmelCase , num_warmup_steps=100 , num_training_steps=(len(__lowerCAmelCase ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase_ = accelerator.prepare( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # Now we train the model for epoch in range(__lowerCAmelCase ): model.train() for step, batch in enumerate(__lowerCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) UpperCAmelCase_ = model(**__lowerCAmelCase ) UpperCAmelCase_ = outputs.loss UpperCAmelCase_ = loss / gradient_accumulation_steps accelerator.backward(__lowerCAmelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() UpperCAmelCase_ = 0 for step, batch in enumerate(__lowerCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase_ = model(**__lowerCAmelCase ) UpperCAmelCase_ = outputs.logits.argmax(dim=-1 ) UpperCAmelCase_ = accelerator.gather((predictions, batch['''labels''']) ) # New Code # # First we check if it's a distributed system if accelerator.use_distributed: # Then see if we're on the last batch of our eval dataloader if step == len(__lowerCAmelCase ) - 1: # Last batch needs to be truncated on distributed systems as it contains additional samples UpperCAmelCase_ = predictions[: len(eval_dataloader.dataset ) - samples_seen] UpperCAmelCase_ = references[: len(eval_dataloader.dataset ) - samples_seen] else: # Otherwise we add the number of samples seen samples_seen += references.shape[0] # All of this can be avoided if you use `Accelerator.gather_for_metrics` instead of `Accelerator.gather`: # accelerator.gather_for_metrics((predictions, batch["labels"])) metric.add_batch( predictions=__lowerCAmelCase , references=__lowerCAmelCase , ) UpperCAmelCase_ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , __lowerCAmelCase ) def lowerCamelCase__ ( ): '''simple docstring''' UpperCAmelCase_ = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__lowerCAmelCase , default=__lowerCAmelCase , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) UpperCAmelCase_ = parser.parse_args() UpperCAmelCase_ = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__lowerCAmelCase , __lowerCAmelCase ) if __name__ == "__main__": main()
350
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) __a = logging.getLogger() def lowerCamelCase__ ( ): '''simple docstring''' UpperCAmelCase_ : List[str] = argparse.ArgumentParser() parser.add_argument('''-f''' ) UpperCAmelCase_ : Dict = parser.parse_args() return args.f class __a( _a ): """simple docstring""" def a__ ( self ) -> None: UpperCAmelCase_ : int = logging.StreamHandler(sys.stdout ) logger.addHandler(_SCREAMING_SNAKE_CASE ) def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> str: UpperCAmelCase_ : int = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 ,'''run_glue_deebert.py''' ) with patch.object(_SCREAMING_SNAKE_CASE ,'''argv''' ,_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : List[str] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(_SCREAMING_SNAKE_CASE ,0.6_66 ) @slow @require_torch_non_multi_gpu def a__ ( self ) -> List[str]: UpperCAmelCase_ : List[Any] = ''' --model_type roberta --model_name_or_path roberta-base --task_name MRPC --do_train --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --max_seq_length 128 --per_gpu_eval_batch_size=1 --per_gpu_train_batch_size=8 --learning_rate 2e-4 --num_train_epochs 3 --overwrite_output_dir --seed 42 --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --save_steps 0 --overwrite_cache --eval_after_first_stage '''.split() self.run_and_check(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Union[str, Any] = ''' --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --eval_each_highway --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 '''.split() self.run_and_check(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Any = ''' --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --early_exit_entropy 0.1 --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 '''.split() self.run_and_check(_SCREAMING_SNAKE_CASE )
235
0
'''simple docstring''' import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer A_ = logging.get_logger(__name__) A_ = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} A_ = { "vocab_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } A_ = { "vocab_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json" ), }, } A_ = { "vocab_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json" ), }, } A_ = { "facebook/dpr-ctx_encoder-single-nq-base": 5_12, "facebook/dpr-ctx_encoder-multiset-base": 5_12, } A_ = { "facebook/dpr-question_encoder-single-nq-base": 5_12, "facebook/dpr-question_encoder-multiset-base": 5_12, } A_ = { "facebook/dpr-reader-single-nq-base": 5_12, "facebook/dpr-reader-multiset-base": 5_12, } A_ = { "facebook/dpr-ctx_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-ctx_encoder-multiset-base": {"do_lower_case": True}, } A_ = { "facebook/dpr-question_encoder-single-nq-base": {"do_lower_case": True}, "facebook/dpr-question_encoder-multiset-base": {"do_lower_case": True}, } A_ = { "facebook/dpr-reader-single-nq-base": {"do_lower_case": True}, "facebook/dpr-reader-multiset-base": {"do_lower_case": True}, } class _snake_case ( __A ): _A : Any = VOCAB_FILES_NAMES _A : int = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP _A : int = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : Tuple = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION _A : Dict = DPRContextEncoderTokenizer class _snake_case ( __A ): _A : Dict = VOCAB_FILES_NAMES _A : str = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP _A : Any = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : List[str] = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION _A : int = DPRQuestionEncoderTokenizer A_ = collections.namedtuple( "DPRSpanPrediction", ["span_score", "relevance_score", "doc_id", "start_index", "end_index", "text"] ) A_ = collections.namedtuple("DPRReaderOutput", ["start_logits", "end_logits", "relevance_logits"]) A_ = R"\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `'tf'`: Return TensorFlow `tf.constant` objects.\n - `'pt'`: Return PyTorch `torch.Tensor` objects.\n - `'np'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer's default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Return:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n " @add_start_docstrings(__A ) class _snake_case : def __call__( self : Any ,SCREAMING_SNAKE_CASE__ : str ,SCREAMING_SNAKE_CASE__ : str = None ,SCREAMING_SNAKE_CASE__ : Any = None ,SCREAMING_SNAKE_CASE__ : str = False ,SCREAMING_SNAKE_CASE__ : List[str] = False ,SCREAMING_SNAKE_CASE__ : Union[str, Any] = None ,SCREAMING_SNAKE_CASE__ : Any = None ,SCREAMING_SNAKE_CASE__ : Union[str, Any] = None ,**SCREAMING_SNAKE_CASE__ : Any ,): if titles is None and texts is None: return super().__call__( __UpperCAmelCase ,padding=__UpperCAmelCase ,truncation=__UpperCAmelCase ,max_length=__UpperCAmelCase ,return_tensors=__UpperCAmelCase ,return_attention_mask=__UpperCAmelCase ,**__UpperCAmelCase ,) elif titles is None or texts is None: SCREAMING_SNAKE_CASE:List[str] = titles if texts is None else texts return super().__call__( __UpperCAmelCase ,__UpperCAmelCase ,padding=__UpperCAmelCase ,truncation=__UpperCAmelCase ,max_length=__UpperCAmelCase ,return_tensors=__UpperCAmelCase ,return_attention_mask=__UpperCAmelCase ,**__UpperCAmelCase ,) SCREAMING_SNAKE_CASE:int = titles if not isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else [titles] SCREAMING_SNAKE_CASE:str = texts if not isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else [texts] SCREAMING_SNAKE_CASE:Optional[int] = len(__UpperCAmelCase ) SCREAMING_SNAKE_CASE:Any = questions if not isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else [questions] * n_passages assert len(__UpperCAmelCase ) == len( __UpperCAmelCase ), F'''There should be as many titles than texts but got {len(__UpperCAmelCase )} titles and {len(__UpperCAmelCase )} texts.''' SCREAMING_SNAKE_CASE:List[Any] = super().__call__(__UpperCAmelCase ,__UpperCAmelCase ,padding=__UpperCAmelCase ,truncation=__UpperCAmelCase )["input_ids"] SCREAMING_SNAKE_CASE:Dict = super().__call__(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ,padding=__UpperCAmelCase ,truncation=__UpperCAmelCase )["input_ids"] SCREAMING_SNAKE_CASE:Optional[int] = { "input_ids": [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(__UpperCAmelCase ,__UpperCAmelCase ) ] } if return_attention_mask is not False: SCREAMING_SNAKE_CASE:List[Any] = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) SCREAMING_SNAKE_CASE:Optional[int] = attention_mask return self.pad(__UpperCAmelCase ,padding=__UpperCAmelCase ,max_length=__UpperCAmelCase ,return_tensors=__UpperCAmelCase ) def __UpperCamelCase ( self : List[Any] ,SCREAMING_SNAKE_CASE__ : Optional[Any] ,SCREAMING_SNAKE_CASE__ : Tuple ,SCREAMING_SNAKE_CASE__ : Dict = 16 ,SCREAMING_SNAKE_CASE__ : List[str] = 64 ,SCREAMING_SNAKE_CASE__ : str = 4 ,): SCREAMING_SNAKE_CASE:Any = reader_input["input_ids"] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE:int = reader_output[:3] SCREAMING_SNAKE_CASE:int = len(__UpperCAmelCase ) SCREAMING_SNAKE_CASE:Union[str, Any] = sorted(range(__UpperCAmelCase ) ,reverse=__UpperCAmelCase ,key=relevance_logits.__getitem__ ) SCREAMING_SNAKE_CASE:Dict = [] for doc_id in sorted_docs: SCREAMING_SNAKE_CASE:List[Any] = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence SCREAMING_SNAKE_CASE:Tuple = sequence_ids.index(self.sep_token_id ,2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: SCREAMING_SNAKE_CASE:Union[str, Any] = sequence_ids.index(self.pad_token_id ) else: SCREAMING_SNAKE_CASE:int = len(__UpperCAmelCase ) SCREAMING_SNAKE_CASE:Union[str, Any] = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] ,end_logits=end_logits[doc_id][passage_offset:sequence_len] ,max_answer_length=__UpperCAmelCase ,top_spans=__UpperCAmelCase ,) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] ,relevance_score=relevance_logits[doc_id] ,doc_id=__UpperCAmelCase ,start_index=__UpperCAmelCase ,end_index=__UpperCAmelCase ,text=self.decode(sequence_ids[start_index : end_index + 1] ) ,) ) if len(__UpperCAmelCase ) >= num_spans: break return nbest_spans_predictions[:num_spans] def __UpperCamelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : List[str] ,SCREAMING_SNAKE_CASE__ : List[Any] ,SCREAMING_SNAKE_CASE__ : Tuple ,SCREAMING_SNAKE_CASE__ : str ,): SCREAMING_SNAKE_CASE:str = [] for start_index, start_score in enumerate(__UpperCAmelCase ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) SCREAMING_SNAKE_CASE:Optional[int] = sorted(__UpperCAmelCase ,key=lambda SCREAMING_SNAKE_CASE__ : x[1] ,reverse=__UpperCAmelCase ) SCREAMING_SNAKE_CASE:int = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F'''Wrong span indices: [{start_index}:{end_index}]''' SCREAMING_SNAKE_CASE:Optional[Any] = end_index - start_index + 1 assert length <= max_answer_length, F'''Span is too long: {length} > {max_answer_length}''' if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(__UpperCAmelCase ) == top_spans: break return chosen_span_intervals @add_end_docstrings(__A ) class _snake_case ( __A , __A ): _A : Tuple = VOCAB_FILES_NAMES _A : List[Any] = READER_PRETRAINED_VOCAB_FILES_MAP _A : str = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : List[str] = READER_PRETRAINED_INIT_CONFIGURATION _A : List[Any] = ['input_ids', 'attention_mask'] _A : Optional[int] = DPRReaderTokenizer
139
"""simple docstring""" import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __lowerCamelCase = "\\n@inproceedings{popovic-2015-chrf,\n title = \"chr{F}: character n-gram {F}-score for automatic {MT} evaluation\",\n author = \"Popovi{\'c}, Maja\",\n booktitle = \"Proceedings of the Tenth Workshop on Statistical Machine Translation\",\n month = sep,\n year = \"2015\",\n address = \"Lisbon, Portugal\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/W15-3049\",\n doi = \"10.18653/v1/W15-3049\",\n pages = \"392--395\",\n}\n@inproceedings{popovic-2017-chrf,\n title = \"chr{F}++: words helping character n-grams\",\n author = \"Popovi{\'c}, Maja\",\n booktitle = \"Proceedings of the Second Conference on Machine Translation\",\n month = sep,\n year = \"2017\",\n address = \"Copenhagen, Denmark\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/W17-4770\",\n doi = \"10.18653/v1/W17-4770\",\n pages = \"612--618\",\n}\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" __lowerCamelCase = "\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n" __lowerCamelCase = "\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n 'score' (float): The chrF (chrF++) score,\n 'char_order' (int): The character n-gram order,\n 'word_order' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n 'beta' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {'score': 84.64214891738334, 'char_order': 6, 'word_order': 0, 'beta': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {'score': 82.87263732906315, 'char_order': 6, 'word_order': 2, 'beta': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = [\"The relationship between cats and dogs is not exactly friendly.\", \"a good bookshop is just a genteel black hole that knows how to read.\"]\n >>> reference = [[\"The relationship between dogs and cats is not exactly friendly.\"], [\"A good bookshop is just a genteel Black Hole that knows how to read.\"]]\n >>> chrf = datasets.load_metric(\"chrf\")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {'score': 92.12853119829202, 'char_order': 6, 'word_order': 2, 'beta': 2}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase__( datasets.Metric ): def snake_case__ ( self ) -> Tuple: if version.parse(scb.__version__ ) < version.parse('1.4.12' ): raise ImportWarning( 'To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n' 'You can install it with `pip install "sacrebleu>=1.4.12"`.' ) return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,homepage='https://github.com/mjpost/sacreBLEU#chrf--chrf' ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { 'predictions': datasets.Value('string' ,id='sequence' ), 'references': datasets.Sequence(datasets.Value('string' ,id='sequence' ) ,id='references' ), } ) ,codebase_urls=['https://github.com/mjpost/sacreBLEU#chrf--chrf'] ,reference_urls=[ 'https://github.com/m-popovic/chrF', ] ,) def snake_case__ ( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase = CHRF.CHAR_ORDER ,__UpperCAmelCase = CHRF.WORD_ORDER ,__UpperCAmelCase = CHRF.BETA ,__UpperCAmelCase = False ,__UpperCAmelCase = False ,__UpperCAmelCase = False ,) -> Union[str, Any]: A__ = len(references[0] ) if any(len(__UpperCAmelCase ) != references_per_prediction for refs in references ): raise ValueError('Sacrebleu requires the same number of references for each prediction' ) A__ = [[refs[i] for refs in references] for i in range(__UpperCAmelCase )] A__ = CHRF(__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ) A__ = sb_chrf.corpus_score(__UpperCAmelCase ,__UpperCAmelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
221
0
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') UpperCAmelCase__ : str =logging.getLogger(__name__) @dataclass class __A : __A = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __A = field( default=a , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __A = field( default=a , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __A = field( default=a , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __A = field( default=a , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) __A = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) __A = field( default=a , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class __A : __A = field(default=a , metadata={"""help""": """The input training data file (a text file)."""} ) __A = field( default=a , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) __A = field( default=a , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __A = field( default=a , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __A = field( default=a , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __A = field( default=a , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) __A = field( default=a , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) __A = field( default=a , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def _snake_case ( self ): if self.train_file is not None: lowerCamelCase =self.train_file.split(""".""" )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: lowerCamelCase =self.validation_file.split(""".""" )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class __A : __A = 42 __A = True __A = None __A = None def __call__( self , UpperCAmelCase_ ): lowerCamelCase ="""label""" if """label""" in features[0].keys() else """labels""" lowerCamelCase =[feature.pop(UpperCAmelCase_ ) for feature in features] lowerCamelCase =len(UpperCAmelCase_ ) lowerCamelCase =len(features[0]["""input_ids"""] ) lowerCamelCase =[ [{k: v[i] for k, v in feature.items()} for i in range(UpperCAmelCase_ )] for feature in features ] lowerCamelCase =list(chain(*UpperCAmelCase_ ) ) lowerCamelCase =self.tokenizer.pad( UpperCAmelCase_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" , ) # Un-flatten lowerCamelCase ={k: v.view(UpperCAmelCase_ , UpperCAmelCase_ , -1 ) for k, v in batch.items()} # Add back labels lowerCamelCase =torch.tensor(UpperCAmelCase_ , dtype=torch.intaa ) return batch def _lowercase ( ) -> Union[str, Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCamelCase =HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCamelCase , lowerCamelCase , lowerCamelCase =parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCamelCase , lowerCamelCase , lowerCamelCase =parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_swag""" , _UpperCAmelCase , _UpperCAmelCase ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCamelCase =training_args.get_process_log_level() logger.setLevel(_UpperCAmelCase ) datasets.utils.logging.set_verbosity(_UpperCAmelCase ) transformers.utils.logging.set_verbosity(_UpperCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. lowerCamelCase =None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCamelCase =get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: lowerCamelCase ={} if data_args.train_file is not None: lowerCamelCase =data_args.train_file if data_args.validation_file is not None: lowerCamelCase =data_args.validation_file lowerCamelCase =data_args.train_file.split(""".""" )[-1] lowerCamelCase =load_dataset( _UpperCAmelCase , data_files=_UpperCAmelCase , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. lowerCamelCase =load_dataset( """swag""" , """regular""" , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCamelCase =AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCamelCase =AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCamelCase =AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=_UpperCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. lowerCamelCase =[F"""ending{i}""" for i in range(4 )] lowerCamelCase ="""sent1""" lowerCamelCase ="""sent2""" if data_args.max_seq_length is None: lowerCamelCase =tokenizer.model_max_length if max_seq_length > 10_24: logger.warning( """The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value""" """ of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can""" """ override this default with `--block_size xxx`.""" ) lowerCamelCase =10_24 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" F"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) lowerCamelCase =min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(_UpperCAmelCase ): lowerCamelCase =[[context] * 4 for context in examples[context_name]] lowerCamelCase =examples[question_header_name] lowerCamelCase =[ [F"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(_UpperCAmelCase ) ] # Flatten out lowerCamelCase =list(chain(*_UpperCAmelCase ) ) lowerCamelCase =list(chain(*_UpperCAmelCase ) ) # Tokenize lowerCamelCase =tokenizer( _UpperCAmelCase , _UpperCAmelCase , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase , padding="""max_length""" if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(_UpperCAmelCase ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError("""--do_train requires a train dataset""" ) lowerCamelCase =raw_datasets["""train"""] if data_args.max_train_samples is not None: lowerCamelCase =min(len(_UpperCAmelCase ) , data_args.max_train_samples ) lowerCamelCase =train_dataset.select(range(_UpperCAmelCase ) ) with training_args.main_process_first(desc="""train dataset map pre-processing""" ): lowerCamelCase =train_dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError("""--do_eval requires a validation dataset""" ) lowerCamelCase =raw_datasets["""validation"""] if data_args.max_eval_samples is not None: lowerCamelCase =min(len(_UpperCAmelCase ) , data_args.max_eval_samples ) lowerCamelCase =eval_dataset.select(range(_UpperCAmelCase ) ) with training_args.main_process_first(desc="""validation dataset map pre-processing""" ): lowerCamelCase =eval_dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator lowerCamelCase =( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=_UpperCAmelCase , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(_UpperCAmelCase ): lowerCamelCase , lowerCamelCase =eval_predictions lowerCamelCase =np.argmax(_UpperCAmelCase , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer lowerCamelCase =Trainer( model=_UpperCAmelCase , args=_UpperCAmelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=_UpperCAmelCase , data_collator=_UpperCAmelCase , compute_metrics=_UpperCAmelCase , ) # Training if training_args.do_train: lowerCamelCase =None if training_args.resume_from_checkpoint is not None: lowerCamelCase =training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCamelCase =last_checkpoint lowerCamelCase =trainer.train(resume_from_checkpoint=_UpperCAmelCase ) trainer.save_model() # Saves the tokenizer too for easy upload lowerCamelCase =train_result.metrics lowerCamelCase =( data_args.max_train_samples if data_args.max_train_samples is not None else len(_UpperCAmelCase ) ) lowerCamelCase =min(_UpperCAmelCase , len(_UpperCAmelCase ) ) trainer.log_metrics("""train""" , _UpperCAmelCase ) trainer.save_metrics("""train""" , _UpperCAmelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("""*** Evaluate ***""" ) lowerCamelCase =trainer.evaluate() lowerCamelCase =data_args.max_eval_samples if data_args.max_eval_samples is not None else len(_UpperCAmelCase ) lowerCamelCase =min(_UpperCAmelCase , len(_UpperCAmelCase ) ) trainer.log_metrics("""eval""" , _UpperCAmelCase ) trainer.save_metrics("""eval""" , _UpperCAmelCase ) lowerCamelCase ={ """finetuned_from""": model_args.model_name_or_path, """tasks""": """multiple-choice""", """dataset_tags""": """swag""", """dataset_args""": """regular""", """dataset""": """SWAG""", """language""": """en""", } if training_args.push_to_hub: trainer.push_to_hub(**_UpperCAmelCase ) else: trainer.create_model_card(**_UpperCAmelCase ) def _lowercase ( _UpperCAmelCase ) -> str: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
262
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def _lowercase ( ) -> str: with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(_UpperCAmelCase ): requests.request("""GET""" , """https://huggingface.co""" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("""GET""" , """https://huggingface.co""" , timeout=1.0 ) @pytest.mark.integration def _lowercase ( ) -> Union[str, Any]: with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("""GET""" , """https://huggingface.co""" ) def _lowercase ( ) -> int: with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(_UpperCAmelCase ): http_head("""https://huggingface.co""" )
262
1
"""simple docstring""" import string import numpy def __lowerCamelCase ( __UpperCamelCase , __UpperCamelCase ) -> int: """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , __UpperCamelCase ) class __lowerCamelCase : '''simple docstring''' a_ : List[str] = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) a_ : Tuple = numpy.vectorize(lambda A__ : x % 36 ) a_ : Optional[Any] = numpy.vectorize(A__ ) def __init__( self : Dict , a_ : numpy.ndarray ): lowerCAmelCase_ : Union[str, Any] = self.modulus(a_ ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key lowerCAmelCase_ : Optional[Any] = encrypt_key.shape[0] def lowerCamelCase ( self : List[str] , a_ : str ): return self.key_string.index(a_ ) def lowerCamelCase ( self : int , a_ : int ): return self.key_string[round(a_ )] def lowerCamelCase ( self : List[Any] ): lowerCAmelCase_ : Optional[int] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: lowerCAmelCase_ : Optional[Any] = det % len(self.key_string ) lowerCAmelCase_ : List[Any] = len(self.key_string ) if greatest_common_divisor(a_ , len(self.key_string ) ) != 1: lowerCAmelCase_ : List[Any] = ( f'''determinant modular {req_l} of encryption key({det}) ''' f'''is not co prime w.r.t {req_l}.\nTry another key.''' ) raise ValueError(a_ ) def lowerCamelCase ( self : List[str] , a_ : str ): lowerCAmelCase_ : Any = [char for char in text.upper() if char in self.key_string] lowerCAmelCase_ : Union[str, Any] = chars[-1] while len(a_ ) % self.break_key != 0: chars.append(a_ ) return "".join(a_ ) def lowerCamelCase ( self : Dict , a_ : str ): lowerCAmelCase_ : Tuple = self.process_text(text.upper() ) lowerCAmelCase_ : Any = "" for i in range(0 , len(a_ ) - self.break_key + 1 , self.break_key ): lowerCAmelCase_ : List[str] = text[i : i + self.break_key] lowerCAmelCase_ : List[str] = [self.replace_letters(a_ ) for char in batch] lowerCAmelCase_ : Any = numpy.array([vec] ).T lowerCAmelCase_ : str = self.modulus(self.encrypt_key.dot(a_ ) ).T.tolist()[ 0 ] lowerCAmelCase_ : List[str] = "".join( self.replace_digits(a_ ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def lowerCamelCase ( self : Optional[Any] ): lowerCAmelCase_ : int = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: lowerCAmelCase_ : Optional[int] = det % len(self.key_string ) lowerCAmelCase_ : Optional[Any] = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: lowerCAmelCase_ : Any = i break lowerCAmelCase_ : Dict = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(a_ ) ) def lowerCamelCase ( self : Dict , a_ : str ): lowerCAmelCase_ : str = self.make_decrypt_key() lowerCAmelCase_ : str = self.process_text(text.upper() ) lowerCAmelCase_ : Union[str, Any] = "" for i in range(0 , len(a_ ) - self.break_key + 1 , self.break_key ): lowerCAmelCase_ : Optional[int] = text[i : i + self.break_key] lowerCAmelCase_ : Optional[Any] = [self.replace_letters(a_ ) for char in batch] lowerCAmelCase_ : Union[str, Any] = numpy.array([vec] ).T lowerCAmelCase_ : Dict = self.modulus(decrypt_key.dot(a_ ) ).T.tolist()[0] lowerCAmelCase_ : Optional[Any] = "".join( self.replace_digits(a_ ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def __lowerCamelCase ( ) -> None: """simple docstring""" lowerCAmelCase_ : List[str] = int(input("Enter the order of the encryption key: " ) ) lowerCAmelCase_ : List[str] = [] print("Enter each row of the encryption key with space separated integers" ) for _ in range(__UpperCamelCase ): lowerCAmelCase_ : Tuple = [int(__UpperCamelCase ) for x in input().split()] hill_matrix.append(__UpperCamelCase ) lowerCAmelCase_ : int = HillCipher(numpy.array(__UpperCamelCase ) ) print("Would you like to encrypt or decrypt some text? (1 or 2)" ) lowerCAmelCase_ : Dict = input("\n1. Encrypt\n2. Decrypt\n" ) if option == "1": lowerCAmelCase_ : List[Any] = input("What text would you like to encrypt?: " ) print("Your encrypted text is:" ) print(hc.encrypt(__UpperCamelCase ) ) elif option == "2": lowerCAmelCase_ : List[Any] = input("What text would you like to decrypt?: " ) print("Your decrypted text is:" ) print(hc.decrypt(__UpperCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
241
"""simple docstring""" # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys lowercase__ = """3""" print("""Python version:""", sys.version) print("""OS platform:""", platform.platform()) print("""OS architecture:""", platform.machine()) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) print("""Cuda version:""", torch.version.cuda) print("""CuDNN version:""", torch.backends.cudnn.version()) print("""Number of GPUs available:""", torch.cuda.device_count()) except ImportError: print("""Torch version:""", None) try: import transformers print("""transformers version:""", transformers.__version__) except ImportError: print("""transformers version:""", None)
241
1
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BlipaProcessor, BlipImageProcessor, GPTaTokenizer, PreTrainedTokenizerFast @require_vision class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =tempfile.mkdtemp() __lowercase =BlipImageProcessor() __lowercase =GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model') __lowercase =BlipaProcessor(__A , __A) processor.save_pretrained(self.tmpdirname) def __lowerCamelCase ( self : Dict , **_lowerCAmelCase : str): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **__A).tokenizer def __lowerCamelCase ( self : Tuple , **_lowerCAmelCase : int): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **__A).image_processor def __lowerCamelCase ( self : int): '''simple docstring''' shutil.rmtree(self.tmpdirname) def __lowerCamelCase ( self : int): '''simple docstring''' __lowercase =[np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta)] __lowercase =[Image.fromarray(np.moveaxis(__A , 0 , -1)) for x in image_inputs] return image_inputs def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __lowercase =BlipaProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) __lowercase =self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)') __lowercase =self.get_image_processor(do_normalize=__A , padding_value=1.0) __lowercase =BlipaProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=__A , padding_value=1.0) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.tokenizer , __A) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , __A) def __lowerCamelCase ( self : Dict): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_tokenizer() __lowercase =BlipaProcessor(tokenizer=__A , image_processor=__A) __lowercase =self.prepare_image_inputs() __lowercase =image_processor(__A , return_tensors='np') __lowercase =processor(images=__A , return_tensors='np') for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) def __lowerCamelCase ( self : Any): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_tokenizer() __lowercase =BlipaProcessor(tokenizer=__A , image_processor=__A) __lowercase ='''lower newer''' __lowercase =processor(text=__A) __lowercase =tokenizer(__A , return_token_type_ids=__A) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_tokenizer() __lowercase =BlipaProcessor(tokenizer=__A , image_processor=__A) __lowercase ='''lower newer''' __lowercase =self.prepare_image_inputs() __lowercase =processor(text=__A , images=__A) self.assertListEqual(list(inputs.keys()) , ['pixel_values', 'input_ids', 'attention_mask']) # test if it raises when no input is passed with pytest.raises(__A): processor() def __lowerCamelCase ( self : Tuple): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_tokenizer() __lowercase =BlipaProcessor(tokenizer=__A , image_processor=__A) __lowercase =[[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowercase =processor.batch_decode(__A) __lowercase =tokenizer.batch_decode(__A) self.assertListEqual(__A , __A) def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_tokenizer() __lowercase =BlipaProcessor(tokenizer=__A , image_processor=__A) __lowercase ='''lower newer''' __lowercase =self.prepare_image_inputs() __lowercase =processor(text=__A , images=__A) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys()) , ['pixel_values', 'input_ids', 'attention_mask'])
360
'''simple docstring''' import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class _UpperCamelCase ( A , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = XLMTokenizer lowerCAmelCase__ = False def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowercase =[ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] __lowercase =dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase)))) __lowercase =['l o 123', 'lo w 1456', 'e r</w> 1789', ''] __lowercase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) __lowercase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file']) with open(self.vocab_file , 'w') as fp: fp.write(json.dumps(_lowerCAmelCase)) with open(self.merges_file , 'w') as fp: fp.write('\n'.join(_lowerCAmelCase)) def __lowerCamelCase ( self : List[str] , _lowerCAmelCase : Any): '''simple docstring''' __lowercase ='lower newer' __lowercase ='lower newer' return input_text, output_text def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =XLMTokenizer(self.vocab_file , self.merges_file) __lowercase ='lower' __lowercase =['low', 'er</w>'] __lowercase =tokenizer.tokenize(_lowerCAmelCase) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase) __lowercase =tokens + ['<unk>'] __lowercase =[1_4, 1_5, 2_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase) , _lowerCAmelCase) @slow def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =XLMTokenizer.from_pretrained('xlm-mlm-en-2048') __lowercase =tokenizer.encode('sequence builders' , add_special_tokens=_lowerCAmelCase) __lowercase =tokenizer.encode('multi-sequence build' , add_special_tokens=_lowerCAmelCase) __lowercase =tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase) __lowercase =tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
48
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""image_processor""", """tokenizer"""] SCREAMING_SNAKE_CASE__ : List[Any] = """ViTImageProcessor""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self , __lowercase=None , __lowercase=None , **__lowercase ) -> int: lowerCAmelCase_ : Optional[Any] = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __lowercase , ) lowerCAmelCase_ : int = kwargs.pop('''feature_extractor''' ) lowerCAmelCase_ : List[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(__lowercase , __lowercase ) def __call__( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase=None , **__lowercase ) -> List[Any]: if text is None and visual_prompt is None and images is None: raise ValueError('''You have to specify either text, visual prompt or images.''' ) if text is not None and visual_prompt is not None: raise ValueError('''You have to specify exactly one type of prompt. Either text or visual prompt.''' ) if text is not None: lowerCAmelCase_ : Tuple = self.tokenizer(__lowercase , return_tensors=__lowercase , **__lowercase ) if visual_prompt is not None: lowerCAmelCase_ : Any = self.image_processor(__lowercase , return_tensors=__lowercase , **__lowercase ) if images is not None: lowerCAmelCase_ : Union[str, Any] = self.image_processor(__lowercase , return_tensors=__lowercase , **__lowercase ) if visual_prompt is not None and images is not None: lowerCAmelCase_ : str = { '''pixel_values''': image_features.pixel_values, '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: lowerCAmelCase_ : Dict = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: lowerCAmelCase_ : int = { '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**__lowercase ) , tensor_type=__lowercase ) def lowercase_ ( self , *__lowercase , **__lowercase ) -> List[str]: return self.tokenizer.batch_decode(*__lowercase , **__lowercase ) def lowercase_ ( self , *__lowercase , **__lowercase ) -> int: return self.tokenizer.decode(*__lowercase , **__lowercase ) @property def lowercase_ ( self ) -> List[Any]: warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __lowercase , ) return self.image_processor_class @property def lowercase_ ( self ) -> List[Any]: warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __lowercase , ) return self.image_processor
262
import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def lowerCAmelCase ( lowerCAmelCase_ )-> Union[str, Any]: return EnvironmentCommand() class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @staticmethod def lowercase_ ( __lowercase ) -> List[Any]: lowerCAmelCase_ : List[str] = parser.add_parser('''env''' ) download_parser.set_defaults(func=__lowercase ) def lowercase_ ( self ) -> int: lowerCAmelCase_ : Optional[Any] = huggingface_hub.__version__ lowerCAmelCase_ : str = '''not installed''' lowerCAmelCase_ : str = '''NA''' if is_torch_available(): import torch lowerCAmelCase_ : Any = torch.__version__ lowerCAmelCase_ : str = torch.cuda.is_available() lowerCAmelCase_ : List[str] = '''not installed''' if is_transformers_available(): import transformers lowerCAmelCase_ : Any = transformers.__version__ lowerCAmelCase_ : Optional[Any] = '''not installed''' if is_accelerate_available(): import accelerate lowerCAmelCase_ : List[Any] = accelerate.__version__ lowerCAmelCase_ : List[str] = '''not installed''' if is_xformers_available(): import xformers lowerCAmelCase_ : Optional[Any] = xformers.__version__ lowerCAmelCase_ : int = { '''`diffusers` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''PyTorch version (GPU?)''': f"""{pt_version} ({pt_cuda_available})""", '''Huggingface_hub version''': hub_version, '''Transformers version''': transformers_version, '''Accelerate version''': accelerate_version, '''xFormers version''': xformers_version, '''Using GPU in script?''': '''<fill in>''', '''Using distributed or parallel set-up in script?''': '''<fill in>''', } print('''\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n''' ) print(self.format_dict(__lowercase ) ) return info @staticmethod def lowercase_ ( __lowercase ) -> str: return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
262
1
"""simple docstring""" import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class snake_case ( __lowercase ): __magic_name__ = (KDPMaDiscreteScheduler,) __magic_name__ = 10 def lowerCamelCase__ ( self : int , **A : List[str] ): '''simple docstring''' a : Any = { '''num_train_timesteps''': 1_1_0_0, '''beta_start''': 0.00_01, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', } config.update(**_a ) return config def lowerCamelCase__ ( self : Dict ): '''simple docstring''' for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=_a ) def lowerCamelCase__ ( self : Tuple ): '''simple docstring''' for beta_start, beta_end in zip([0.0_00_01, 0.00_01, 0.0_01] , [0.00_02, 0.0_02, 0.02] ): self.check_over_configs(beta_start=_a , beta_end=_a ) def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=_a ) def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_a ) def lowerCamelCase__ ( self : str ): '''simple docstring''' a : int = self.scheduler_classes[0] a : Dict = self.get_scheduler_config(prediction_type='v_prediction' ) a : List[str] = scheduler_class(**_a ) scheduler.set_timesteps(self.num_inference_steps ) a : Optional[Any] = self.dummy_model() a : str = self.dummy_sample_deter * scheduler.init_noise_sigma a : Dict = sample.to(_a ) for i, t in enumerate(scheduler.timesteps ): a : Tuple = scheduler.scale_model_input(_a , _a ) a : Dict = model(_a , _a ) a : str = scheduler.step(_a , _a , _a ) a : Dict = output.prev_sample a : Optional[Any] = torch.sum(torch.abs(_a ) ) a : Any = torch.mean(torch.abs(_a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693_4286_5017_0972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.00_02 ) < 1E-3 def lowerCamelCase__ ( self : Optional[int] ): '''simple docstring''' if torch_device == "mps": return a : Optional[int] = self.scheduler_classes[0] a : int = self.get_scheduler_config() a : Optional[int] = scheduler_class(**_a ) scheduler.set_timesteps(self.num_inference_steps ) a : Any = self.dummy_model() a : Any = self.dummy_sample_deter * scheduler.init_noise_sigma a : Optional[Any] = sample.to(_a ) for i, t in enumerate(scheduler.timesteps ): a : Optional[Any] = scheduler.scale_model_input(_a , _a ) a : List[Any] = model(_a , _a ) a : Optional[Any] = scheduler.step(_a , _a , _a ) a : Optional[Any] = output.prev_sample a : Dict = torch.sum(torch.abs(_a ) ) a : Tuple = torch.mean(torch.abs(_a ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 def lowerCamelCase__ ( self : Any ): '''simple docstring''' if torch_device == "mps": return a : List[str] = self.scheduler_classes[0] a : str = self.get_scheduler_config() a : List[Any] = scheduler_class(**_a ) scheduler.set_timesteps(self.num_inference_steps , device=_a ) a : Union[str, Any] = self.dummy_model() a : Tuple = self.dummy_sample_deter.to(_a ) * scheduler.init_noise_sigma for t in scheduler.timesteps: a : str = scheduler.scale_model_input(_a , _a ) a : int = model(_a , _a ) a : int = scheduler.step(_a , _a , _a ) a : Dict = output.prev_sample a : Tuple = torch.sum(torch.abs(_a ) ) a : Union[str, Any] = torch.mean(torch.abs(_a ) ) if str(_a ).startswith('cpu' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3
365
"""simple docstring""" def snake_case (A_ :int ): '''simple docstring''' if isinstance(A_ , A_ ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if isinstance(A_ , A_ ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if num == 0: return "0b0" a : List[Any] = False if num < 0: a : Optional[int] = True a : Dict = -num a : list[int] = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(A_ ) for e in binary ) return "0b" + "".join(str(A_ ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
186
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase_ : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[int] = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Dict = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys UpperCAmelCase_ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
38
def __UpperCAmelCase ( __a : str ) -> int: """simple docstring""" assert column_title.isupper() _a : Optional[Any] = 0 _a : List[Any] = len(__a ) - 1 _a : List[str] = 0 while index >= 0: _a : Dict = (ord(column_title[index] ) - 64) * pow(26 ,__a ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
235
0
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( __lowerCamelCase ): '''simple docstring''' _lowerCamelCase = ['image_processor', 'tokenizer'] _lowerCamelCase = 'ViTImageProcessor' _lowerCamelCase = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__( self ,lowerCamelCase_=None ,lowerCamelCase_=None ,**lowerCamelCase_ ) -> str: A = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" ,UpperCamelCase_ ,) A = kwargs.pop("""feature_extractor""" ) A = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(UpperCamelCase_ ,UpperCamelCase_ ) def __call__( self ,lowerCamelCase_=None ,lowerCamelCase_=None ,lowerCamelCase_=None ,lowerCamelCase_=None ,**lowerCamelCase_ ) -> Union[str, Any]: if text is None and visual_prompt is None and images is None: raise ValueError("""You have to specify either text, visual prompt or images.""" ) if text is not None and visual_prompt is not None: raise ValueError("""You have to specify exactly one type of prompt. Either text or visual prompt.""" ) if text is not None: A = self.tokenizer(UpperCamelCase_ ,return_tensors=UpperCamelCase_ ,**UpperCamelCase_ ) if visual_prompt is not None: A = self.image_processor(UpperCamelCase_ ,return_tensors=UpperCamelCase_ ,**UpperCamelCase_ ) if images is not None: A = self.image_processor(UpperCamelCase_ ,return_tensors=UpperCamelCase_ ,**UpperCamelCase_ ) if visual_prompt is not None and images is not None: A = { """pixel_values""": image_features.pixel_values, """conditional_pixel_values""": prompt_features.pixel_values, } return encoding elif text is not None and images is not None: A = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: A = { """conditional_pixel_values""": prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**UpperCamelCase_ ) ,tensor_type=UpperCamelCase_ ) def UpperCamelCase__ ( self ,*lowerCamelCase_ ,**lowerCamelCase_ ) -> Optional[int]: return self.tokenizer.batch_decode(*UpperCamelCase_ ,**UpperCamelCase_ ) def UpperCamelCase__ ( self ,*lowerCamelCase_ ,**lowerCamelCase_ ) -> Optional[int]: return self.tokenizer.decode(*UpperCamelCase_ ,**UpperCamelCase_ ) @property def UpperCamelCase__ ( self ) -> List[str]: warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" ,UpperCamelCase_ ,) return self.image_processor_class @property def UpperCamelCase__ ( self ) -> Union[str, Any]: warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" ,UpperCamelCase_ ,) return self.image_processor
357
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase__ : '''simple docstring''' def __init__( self ,lowerCamelCase_ ,lowerCamelCase_=1_2 ,lowerCamelCase_=7 ,lowerCamelCase_=True ,lowerCamelCase_=True ,lowerCamelCase_=True ,lowerCamelCase_=9_9 ,lowerCamelCase_=3_2 ,lowerCamelCase_=3_2 ,lowerCamelCase_=2 ,lowerCamelCase_=4 ,lowerCamelCase_=3_7 ,lowerCamelCase_=0.1 ,lowerCamelCase_=0.1 ,lowerCamelCase_=5_1_2 ,lowerCamelCase_=0.02 ,lowerCamelCase_=0 ,lowerCamelCase_=None ,) -> List[str]: A = parent A = batch_size A = seq_length A = is_training A = use_input_mask A = use_labels A = vocab_size A = hidden_size A = projection_dim A = num_hidden_layers A = num_attention_heads A = intermediate_size A = dropout A = attention_dropout A = max_position_embeddings A = initializer_range A = scope A = bos_token_id def UpperCamelCase__ ( self ) -> Tuple: A = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) A = None if self.use_input_mask: A = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: A = input_mask.numpy() A , A = input_mask.shape A = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCamelCase_ ): A = 1 A = 0 A = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCamelCase_ ) def UpperCamelCase__ ( self ) -> int: return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) -> Tuple: A = TFBlipTextModel(config=lowerCamelCase_ ) A = model(lowerCamelCase_ ,attention_mask=lowerCamelCase_ ,training=lowerCamelCase_ ) A = model(lowerCamelCase_ ,training=lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCamelCase__ ( self ) -> Optional[Any]: A = self.prepare_config_and_inputs() A , A , A = config_and_inputs A = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class lowerCamelCase__ ( SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' _lowerCamelCase = (TFBlipTextModel,) if is_tf_available() else () _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def UpperCamelCase__ ( self ) -> List[str]: A = BlipTextModelTester(self ) A = ConfigTester(self ,config_class=lowerCamelCase_ ,hidden_size=3_7 ) def UpperCamelCase__ ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() def UpperCamelCase__ ( self ) -> Union[str, Any]: A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def UpperCamelCase__ ( self ) -> Optional[int]: pass def UpperCamelCase__ ( self ) -> Optional[Any]: pass @unittest.skip(reason="""Blip does not use inputs_embeds""" ) def UpperCamelCase__ ( self ) -> Optional[int]: pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def UpperCamelCase__ ( self ) -> Dict: pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def UpperCamelCase__ ( self ) -> str: pass @slow def UpperCamelCase__ ( self ) -> str: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A = TFBlipTextModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_=True ) -> str: super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCamelCase_ )
77
0
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging _UpperCAmelCase : List[str] =logging.get_logger(__name__) _UpperCAmelCase : List[Any] ={ """google/umt5-small""": """https://huggingface.co/google/umt5-small/resolve/main/config.json""", # See all umt5 models at https://huggingface.co/models?filter=umt5 } class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = """umt5""" SCREAMING_SNAKE_CASE__ : str = ["""past_key_values"""] def __init__( self , __lowercase=2_5_0_1_1_2 , __lowercase=5_1_2 , __lowercase=6_4 , __lowercase=1_0_2_4 , __lowercase=8 , __lowercase=None , __lowercase=6 , __lowercase=3_2 , __lowercase=1_2_8 , __lowercase=0.1 , __lowercase=1e-6 , __lowercase=1.0 , __lowercase="gated-gelu" , __lowercase=True , __lowercase=True , __lowercase="T5Tokenizer" , __lowercase=True , __lowercase=0 , __lowercase=1 , __lowercase=0 , **__lowercase , ) -> str: super().__init__( is_encoder_decoder=__lowercase , tokenizer_class=__lowercase , tie_word_embeddings=__lowercase , pad_token_id=__lowercase , eos_token_id=__lowercase , decoder_start_token_id=__lowercase , **__lowercase , ) lowerCAmelCase_ : Any = vocab_size lowerCAmelCase_ : Optional[Any] = d_model lowerCAmelCase_ : Tuple = d_kv lowerCAmelCase_ : List[str] = d_ff lowerCAmelCase_ : Union[str, Any] = num_layers lowerCAmelCase_ : Dict = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry lowerCAmelCase_ : Dict = num_heads lowerCAmelCase_ : int = relative_attention_num_buckets lowerCAmelCase_ : str = relative_attention_max_distance lowerCAmelCase_ : int = dropout_rate lowerCAmelCase_ : Optional[Any] = layer_norm_epsilon lowerCAmelCase_ : Tuple = initializer_factor lowerCAmelCase_ : List[str] = feed_forward_proj lowerCAmelCase_ : str = use_cache lowerCAmelCase_ : List[str] = self.feed_forward_proj.split('''-''' ) lowerCAmelCase_ : int = act_info[-1] lowerCAmelCase_ : List[Any] = act_info[0] == '''gated''' if len(__lowercase ) > 1 and act_info[0] != "gated" or len(__lowercase ) > 2: raise ValueError( f"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" '''Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ''' '''\'gated-gelu\' or \'relu\'''' ) if feed_forward_proj == "gated-gelu": lowerCAmelCase_ : int = '''gelu_new''' @property def lowercase_ ( self ) -> List[str]: return self.d_model @property def lowercase_ ( self ) -> Optional[Any]: return self.num_heads @property def lowercase_ ( self ) -> Optional[int]: return self.num_layers class snake_case__( UpperCAmelCase__ ): '''simple docstring''' @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.inputs def lowercase_ ( self ) -> Mapping[str, Mapping[int, str]]: lowerCAmelCase_ : List[str] = { '''input_ids''': {0: '''batch''', 1: '''encoder_sequence'''}, '''attention_mask''': {0: '''batch''', 1: '''encoder_sequence'''}, } if self.use_past: lowerCAmelCase_ : List[str] = '''past_encoder_sequence + sequence''' lowerCAmelCase_ : int = {0: '''batch'''} lowerCAmelCase_ : List[str] = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: lowerCAmelCase_ : str = {0: '''batch''', 1: '''decoder_sequence'''} lowerCAmelCase_ : int = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(__lowercase , direction='''inputs''' ) return common_inputs @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.default_onnx_opset def lowercase_ ( self ) -> int: return 1_3 @property def lowercase_ ( self ) -> float: return 5e-4
262
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor _UpperCAmelCase : Any =logging.get_logger(__name__) class snake_case__( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *__lowercase , **__lowercase ) -> None: warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , __lowercase , ) super().__init__(*__lowercase , **__lowercase )
262
1
from __future__ import annotations def lowerCamelCase_ ( _a , _a , _a , _a ): # noqa: E741 """simple docstring""" while r - l > 1: lowerCAmelCase__ : Dict = (l + r) // 2 if v[m] >= key: lowerCAmelCase__ : List[Any] = m else: lowerCAmelCase__ : Union[str, Any] = m # noqa: E741 return r def lowerCamelCase_ ( _a ): """simple docstring""" if len(_UpperCamelCase ) == 0: return 0 lowerCAmelCase__ : List[Any] = [0] * len(_UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = 1 lowerCAmelCase__ : List[str] = v[0] for i in range(1 , len(_UpperCamelCase ) ): if v[i] < tail[0]: lowerCAmelCase__ : Optional[int] = v[i] elif v[i] > tail[length - 1]: lowerCAmelCase__ : Optional[Any] = v[i] length += 1 else: lowerCAmelCase__ : Union[str, Any] = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
359
from __future__ import annotations def lowerCamelCase_ ( _a , _a , _a , _a ): # noqa: E741 """simple docstring""" while r - l > 1: lowerCAmelCase__ : Any = (l + r) // 2 if v[m] >= key: lowerCAmelCase__ : int = m else: lowerCAmelCase__ : Tuple = m # noqa: E741 return r def lowerCamelCase_ ( _a ): """simple docstring""" if len(_a ) == 0: return 0 lowerCAmelCase__ : Optional[int] = [0] * len(_a ) lowerCAmelCase__ : List[Any] = 1 lowerCAmelCase__ : int = v[0] for i in range(1 , len(_a ) ): if v[i] < tail[0]: lowerCAmelCase__ : str = v[i] elif v[i] > tail[length - 1]: lowerCAmelCase__ : Any = v[i] length += 1 else: lowerCAmelCase__ : int = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
211
0
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() __snake_case = logging.get_logger('''transformers.models.speecht5''') def a ( __a , __a , __a ) -> str: '''simple docstring''' hf_model.apply_weight_norm() UpperCamelCase__ :List[str] = checkpoint['''input_conv.weight_g'''] UpperCamelCase__ :Any = checkpoint['''input_conv.weight_v'''] UpperCamelCase__ :Dict = checkpoint['''input_conv.bias'''] for i in range(len(config.upsample_rates ) ): UpperCamelCase__ :Optional[Any] = checkpoint[f'''upsamples.{i}.1.weight_g'''] UpperCamelCase__ :List[str] = checkpoint[f'''upsamples.{i}.1.weight_v'''] UpperCamelCase__ :Dict = checkpoint[f'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCamelCase__ :int = checkpoint[f'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCamelCase__ :Dict = checkpoint[f'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCamelCase__ :Any = checkpoint[f'''blocks.{i}.convs1.{j}.1.bias'''] UpperCamelCase__ :Union[str, Any] = checkpoint[f'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCamelCase__ :Optional[Any] = checkpoint[f'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCamelCase__ :str = checkpoint[f'''blocks.{i}.convs2.{j}.1.bias'''] UpperCamelCase__ :Tuple = checkpoint['''output_conv.1.weight_g'''] UpperCamelCase__ :Dict = checkpoint['''output_conv.1.weight_v'''] UpperCamelCase__ :List[str] = checkpoint['''output_conv.1.bias'''] hf_model.remove_weight_norm() @torch.no_grad() def a ( __a , __a , __a , __a=None , __a=None , ) -> str: '''simple docstring''' if config_path is not None: UpperCamelCase__ :int = SpeechTaHifiGanConfig.from_pretrained(__a ) else: UpperCamelCase__ :int = SpeechTaHifiGanConfig() UpperCamelCase__ :Any = SpeechTaHifiGan(__a ) UpperCamelCase__ :Tuple = torch.load(__a ) load_weights(orig_checkpoint['''model''']['''generator'''] , __a , __a ) UpperCamelCase__ :Optional[int] = np.load(__a ) UpperCamelCase__ :int = stats[0].reshape(-1 ) UpperCamelCase__ :Optional[int] = stats[1].reshape(-1 ) UpperCamelCase__ :str = torch.from_numpy(__a ).float() UpperCamelCase__ :List[Any] = torch.from_numpy(__a ).float() model.save_pretrained(__a ) if repo_id: print('''Pushing to the hub...''' ) model.push_to_hub(__a ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--stats_path''', required=True, default=None, type=str, help='''Path to stats.npy file''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) __snake_case = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
97
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE__ : Optional[int] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Dict = { 'salesforce/blip2-opt-2.7b': 'https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json', } class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : Union[str, Any] = """blip_2_vision_model""" def __init__( self , UpperCamelCase__=1408 , UpperCamelCase__=6144 , UpperCamelCase__=39 , UpperCamelCase__=16 , UpperCamelCase__=224 , UpperCamelCase__=14 , UpperCamelCase__="gelu" , UpperCamelCase__=0.00001 , UpperCamelCase__=0.0 , UpperCamelCase__=1e-10 , UpperCamelCase__=True , **UpperCamelCase__ , ) -> Optional[Any]: super().__init__(**UpperCamelCase__ ) lowerCamelCase : Dict = hidden_size lowerCamelCase : Union[str, Any] = intermediate_size lowerCamelCase : List[str] = num_hidden_layers lowerCamelCase : List[str] = num_attention_heads lowerCamelCase : Dict = patch_size lowerCamelCase : Tuple = image_size lowerCamelCase : Dict = initializer_range lowerCamelCase : Union[str, Any] = attention_dropout lowerCamelCase : Dict = layer_norm_eps lowerCamelCase : Optional[Any] = hidden_act lowerCamelCase : str = qkv_bias @classmethod def _lowercase ( cls , UpperCamelCase__ , **UpperCamelCase__ ) -> "PretrainedConfig": cls._set_token_in_kwargs(UpperCamelCase__ ) lowerCamelCase , lowerCamelCase : List[str] = cls.get_config_dict(UpperCamelCase__ , **UpperCamelCase__ ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": lowerCamelCase : Optional[int] = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(UpperCamelCase__ , **UpperCamelCase__ ) class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : Dict = """blip_2_qformer""" def __init__( self , UpperCamelCase__=3_0522 , UpperCamelCase__=768 , UpperCamelCase__=12 , UpperCamelCase__=12 , UpperCamelCase__=3072 , UpperCamelCase__="gelu" , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=512 , UpperCamelCase__=0.02 , UpperCamelCase__=1e-12 , UpperCamelCase__=0 , UpperCamelCase__="absolute" , UpperCamelCase__=2 , UpperCamelCase__=1408 , **UpperCamelCase__ , ) -> int: super().__init__(pad_token_id=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase : Optional[int] = vocab_size lowerCamelCase : int = hidden_size lowerCamelCase : Dict = num_hidden_layers lowerCamelCase : Union[str, Any] = num_attention_heads lowerCamelCase : int = hidden_act lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : Dict = hidden_dropout_prob lowerCamelCase : Dict = attention_probs_dropout_prob lowerCamelCase : Dict = max_position_embeddings lowerCamelCase : List[str] = initializer_range lowerCamelCase : List[str] = layer_norm_eps lowerCamelCase : int = position_embedding_type lowerCamelCase : Tuple = cross_attention_frequency lowerCamelCase : Optional[int] = encoder_hidden_size @classmethod def _lowercase ( cls , UpperCamelCase__ , **UpperCamelCase__ ) -> "PretrainedConfig": cls._set_token_in_kwargs(UpperCamelCase__ ) lowerCamelCase , lowerCamelCase : str = cls.get_config_dict(UpperCamelCase__ , **UpperCamelCase__ ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": lowerCamelCase : int = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(UpperCamelCase__ , **UpperCamelCase__ ) class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : List[str] = """blip-2""" lowerCamelCase_ : int = True def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=32 , **UpperCamelCase__ ) -> str: super().__init__(**UpperCamelCase__ ) if vision_config is None: lowerCamelCase : List[Any] = {} logger.info("vision_config is None. initializing the Blip2VisionConfig with default values." ) if qformer_config is None: lowerCamelCase : List[Any] = {} logger.info("qformer_config is None. Initializing the Blip2QFormerConfig with default values." ) if text_config is None: lowerCamelCase : Any = {} logger.info("text_config is None. Initializing the text config with default values (`OPTConfig`)." ) lowerCamelCase : Optional[int] = BlipaVisionConfig(**UpperCamelCase__ ) lowerCamelCase : str = BlipaQFormerConfig(**UpperCamelCase__ ) lowerCamelCase : List[str] = text_config["model_type"] if "model_type" in text_config else "opt" lowerCamelCase : str = CONFIG_MAPPING[text_model_type](**UpperCamelCase__ ) lowerCamelCase : Optional[Any] = self.text_config.tie_word_embeddings lowerCamelCase : int = self.text_config.is_encoder_decoder lowerCamelCase : Optional[Any] = num_query_tokens lowerCamelCase : int = self.vision_config.hidden_size lowerCamelCase : Tuple = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES lowerCamelCase : Dict = 1.0 lowerCamelCase : List[Any] = 0.02 @classmethod def _lowercase ( cls , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ , ) -> str: return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **UpperCamelCase__ , ) def _lowercase ( self ) -> Optional[Any]: lowerCamelCase : Tuple = copy.deepcopy(self.__dict__ ) lowerCamelCase : Tuple = self.vision_config.to_dict() lowerCamelCase : int = self.qformer_config.to_dict() lowerCamelCase : Optional[Any] = self.text_config.to_dict() lowerCamelCase : int = self.__class__.model_type return output
48
0
import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class lowerCamelCase__( unittest.TestCase): def lowerCAmelCase__ ( self: int ): __lowerCamelCase = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__lowerCamelCase ) ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__lowerCamelCase ) ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(__lowerCamelCase ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__lowerCamelCase ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', # Removed: 'text_encoder/model.safetensors', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertFalse(is_safetensors_compatible(__lowerCamelCase ) ) def lowerCAmelCase__ ( self: Dict ): __lowerCamelCase = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] __lowerCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCamelCase , variant=__lowerCamelCase ) ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = [ '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] __lowerCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCamelCase , variant=__lowerCamelCase ) ) def lowerCAmelCase__ ( self: Optional[int] ): __lowerCamelCase = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] __lowerCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCamelCase , variant=__lowerCamelCase ) ) def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] __lowerCamelCase = '''fp16''' self.assertFalse(is_safetensors_compatible(__lowerCamelCase , variant=__lowerCamelCase ) ) def lowerCAmelCase__ ( self: int ): __lowerCamelCase = [ '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', ] __lowerCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCamelCase , variant=__lowerCamelCase ) ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] __lowerCamelCase = '''fp16''' self.assertTrue(is_safetensors_compatible(__lowerCamelCase , variant=__lowerCamelCase ) ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', # 'text_encoder/model.fp16.safetensors', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] __lowerCamelCase = '''fp16''' self.assertFalse(is_safetensors_compatible(__lowerCamelCase , variant=__lowerCamelCase ) )
355
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Union[str, Any] = 'yolos' def __init__( self: Dict , UpperCamelCase_: List[Any]=7_68 , UpperCamelCase_: Tuple=12 , UpperCamelCase_: int=12 , UpperCamelCase_: int=30_72 , UpperCamelCase_: List[str]="gelu" , UpperCamelCase_: Union[str, Any]=0.0 , UpperCamelCase_: int=0.0 , UpperCamelCase_: Optional[int]=0.02 , UpperCamelCase_: Dict=1E-12 , UpperCamelCase_: List[Any]=[5_12, 8_64] , UpperCamelCase_: Optional[int]=16 , UpperCamelCase_: Any=3 , UpperCamelCase_: Union[str, Any]=True , UpperCamelCase_: List[str]=1_00 , UpperCamelCase_: List[str]=True , UpperCamelCase_: Any=False , UpperCamelCase_: Optional[Any]=1 , UpperCamelCase_: Any=5 , UpperCamelCase_: Any=2 , UpperCamelCase_: Tuple=5 , UpperCamelCase_: str=2 , UpperCamelCase_: Any=0.1 , **UpperCamelCase_: Any , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = image_size __lowerCamelCase = patch_size __lowerCamelCase = num_channels __lowerCamelCase = qkv_bias __lowerCamelCase = num_detection_tokens __lowerCamelCase = use_mid_position_embeddings __lowerCamelCase = auxiliary_loss # Hungarian matcher __lowerCamelCase = class_cost __lowerCamelCase = bbox_cost __lowerCamelCase = giou_cost # Loss coefficients __lowerCamelCase = bbox_loss_coefficient __lowerCamelCase = giou_loss_coefficient __lowerCamelCase = eos_coefficient class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = version.parse('1.11') @property def lowerCAmelCase__ ( self: Any ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCAmelCase__ ( self: Dict ): return 1E-4 @property def lowerCAmelCase__ ( self: Dict ): return 12
29
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ = { 'configuration_blip_2': [ 'BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Blip2Config', 'Blip2QFormerConfig', 'Blip2VisionConfig', ], 'processing_blip_2': ['Blip2Processor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ 'BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST', 'Blip2Model', 'Blip2QFormerModel', 'Blip2PreTrainedModel', 'Blip2ForConditionalGeneration', 'Blip2VisionModel', ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
11
import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCamelCase : """simple docstring""" def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=10 , _SCREAMING_SNAKE_CASE=[10, 20, 30, 40] , _SCREAMING_SNAKE_CASE=[1, 1, 2, 1] , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE="relu" , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=None , )->List[str]: '''simple docstring''' A_ : str = parent A_ : int = batch_size A_ : List[str] = image_size A_ : Dict = num_channels A_ : Tuple = embeddings_size A_ : Union[str, Any] = hidden_sizes A_ : Dict = depths A_ : str = is_training A_ : Union[str, Any] = use_labels A_ : Union[str, Any] = hidden_act A_ : Optional[Any] = num_labels A_ : Tuple = scope A_ : Optional[int] = len(_SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Optional[Any]: '''simple docstring''' A_ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : str = None if self.use_labels: A_ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Optional[Any] = self.get_config() return config, pixel_values, labels def _snake_case ( self )->Union[str, Any]: '''simple docstring''' return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )->Union[str, Any]: '''simple docstring''' A_ : Dict = RegNetModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ : Any = model(_SCREAMING_SNAKE_CASE ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _snake_case ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )->Union[str, Any]: '''simple docstring''' A_ : Union[str, Any] = self.num_labels A_ : Dict = RegNetForImageClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ : int = model(_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self )->Union[str, Any]: '''simple docstring''' A_ : Tuple = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Any = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _lowerCamelCase ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): """simple docstring""" snake_case = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () snake_case = ( {"feature-extraction": RegNetModel, "image-classification": RegNetForImageClassification} if is_torch_available() else {} ) snake_case = False snake_case = False snake_case = False snake_case = False def _snake_case ( self )->Union[str, Any]: '''simple docstring''' A_ : Union[str, Any] = RegNetModelTester(self ) A_ : Union[str, Any] = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Dict: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _snake_case ( self )->Tuple: '''simple docstring''' return @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def _snake_case ( self )->Dict: '''simple docstring''' pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def _snake_case ( self )->str: '''simple docstring''' pass def _snake_case ( self )->List[Any]: '''simple docstring''' A_ , A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : str = model_class(_SCREAMING_SNAKE_CASE ) A_ : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : Any = [*signature.parameters.keys()] A_ : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Any: '''simple docstring''' A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Optional[Any]: '''simple docstring''' A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Union[str, Any] = model_class(config=_SCREAMING_SNAKE_CASE ) for name, module in model.named_modules(): if isinstance(_SCREAMING_SNAKE_CASE , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) def _snake_case ( self )->List[Any]: '''simple docstring''' def check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): A_ : str = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): A_ : Tuple = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) A_ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() A_ : int = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : int = layer_type A_ : List[Any] = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def _snake_case ( self )->Dict: '''simple docstring''' A_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_SCREAMING_SNAKE_CASE ) @slow def _snake_case ( self )->str: '''simple docstring''' for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = RegNetModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def _SCREAMING_SNAKE_CASE ( ): A_ : int = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _lowerCamelCase ( unittest.TestCase ): """simple docstring""" @cached_property def _snake_case ( self )->List[str]: '''simple docstring''' return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _snake_case ( self )->Tuple: '''simple docstring''' A_ : List[Any] = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_SCREAMING_SNAKE_CASE ) A_ : Optional[Any] = self.default_image_processor A_ : Any = prepare_img() A_ : Optional[Any] = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): A_ : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE ) # verify the logits A_ : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) A_ : Optional[int] = torch.tensor([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) )
186
0
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class snake_case__ : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=32 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=[10, 20, 30, 40] , lowerCAmelCase__=[2, 2, 3, 2] , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=37 , lowerCAmelCase__="gelu" , lowerCAmelCase__=10 , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=["stage2", "stage3", "stage4"] , lowerCAmelCase__=3 , lowerCAmelCase__=None , ) -> Optional[int]: __magic_name__ : Any = parent __magic_name__ : Union[str, Any] = batch_size __magic_name__ : int = image_size __magic_name__ : List[Any] = num_channels __magic_name__ : Tuple = num_stages __magic_name__ : Union[str, Any] = hidden_sizes __magic_name__ : int = depths __magic_name__ : Tuple = is_training __magic_name__ : Optional[int] = use_labels __magic_name__ : int = intermediate_size __magic_name__ : Any = hidden_act __magic_name__ : Optional[Any] = type_sequence_label_size __magic_name__ : Optional[int] = initializer_range __magic_name__ : Any = out_features __magic_name__ : List[str] = num_labels __magic_name__ : str = scope __magic_name__ : Union[str, Any] = num_stages def __magic_name__ ( self ) -> int: __magic_name__ : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __magic_name__ : List[Any] = None if self.use_labels: __magic_name__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__ : Optional[int] = self.get_config() return config, pixel_values, labels def __magic_name__ ( self ) -> Optional[int]: return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def __magic_name__ ( self ) -> Optional[int]: return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=5_12 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=__A , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=2_56 , auxiliary_num_convs=1 , auxiliary_concat_input=__A , loss_ignore_index=2_55 , num_labels=self.num_labels , ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: __magic_name__ : Optional[int] = UperNetForSemanticSegmentation(config=__A ) model.to(__A ) model.eval() __magic_name__ : Union[str, Any] = model(__A ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def __magic_name__ ( self ) -> str: __magic_name__ : Dict = self.prepare_config_and_inputs() ( __magic_name__ ) : List[Any] = config_and_inputs __magic_name__ : str = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class snake_case__ ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): lowercase__ : List[str] = (UperNetForSemanticSegmentation,) if is_torch_available() else () lowercase__ : List[Any] = {'''image-segmentation''': UperNetForSemanticSegmentation} if is_torch_available() else {} lowercase__ : int = False lowercase__ : Optional[Any] = False lowercase__ : List[str] = False lowercase__ : Union[str, Any] = False lowercase__ : Optional[Any] = False lowercase__ : Tuple = False def __magic_name__ ( self ) -> str: __magic_name__ : List[str] = UperNetModelTester(self ) __magic_name__ : Tuple = ConfigTester(self , config_class=__A , has_text_modality=__A , hidden_size=37 ) def __magic_name__ ( self ) -> str: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __magic_name__ ( self ) -> Optional[int]: return def __magic_name__ ( self ) -> Union[str, Any]: __magic_name__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __magic_name__ : str = model_class(__A ) __magic_name__ : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __magic_name__ : Optional[Any] = [*signature.parameters.keys()] __magic_name__ : List[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __A ) def __magic_name__ ( self ) -> Any: __magic_name__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__A ) @unittest.skip(reason="""UperNet does not use inputs_embeds""" ) def __magic_name__ ( self ) -> Optional[int]: pass @unittest.skip(reason="""UperNet does not support input and output embeddings""" ) def __magic_name__ ( self ) -> List[str]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def __magic_name__ ( self ) -> Optional[int]: pass @unittest.skip(reason="""UperNet does not have a base model""" ) def __magic_name__ ( self ) -> str: pass @require_torch_multi_gpu @unittest.skip(reason="""UperNet has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`""" ) def __magic_name__ ( self ) -> List[str]: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def __magic_name__ ( self ) -> Union[str, Any]: pass def __magic_name__ ( self ) -> Union[str, Any]: def check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): __magic_name__ : Union[str, Any] = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): __magic_name__ : str = model(**self._prepare_for_class(__A , __A ) ) __magic_name__ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __magic_name__ : Union[str, Any] = self.model_tester.num_stages self.assertEqual(len(__A ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __magic_name__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __magic_name__ : Optional[Any] = True check_hidden_states_output(__A , __A , __A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __magic_name__ : Tuple = True check_hidden_states_output(__A , __A , __A ) def __magic_name__ ( self ) -> List[str]: __magic_name__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ : int = _config_zero_init(__A ) __magic_name__ : Optional[Any] = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: __magic_name__ : Union[str, Any] = model_class(config=__A ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip(reason="""UperNet does not have tied weights""" ) def __magic_name__ ( self ) -> int: pass @slow def __magic_name__ ( self ) -> Dict: for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __magic_name__ : List[Any] = UperNetForSemanticSegmentation.from_pretrained(__A ) self.assertIsNotNone(__A ) def UpperCamelCase ( ): """simple docstring""" __magic_name__ : Tuple = hf_hub_download( repo_id="""hf-internal-testing/fixtures_ade20k""", repo_type="""dataset""", filename="""ADE_val_00000001.jpg""" ) __magic_name__ : str = Image.open(_A ).convert("""RGB""" ) return image @require_torch @require_vision @slow class snake_case__ ( unittest.TestCase ): def __magic_name__ ( self ) -> Any: __magic_name__ : int = AutoImageProcessor.from_pretrained("""openmmlab/upernet-swin-tiny""" ) __magic_name__ : Union[str, Any] = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-swin-tiny""" ).to(__A ) __magic_name__ : Optional[Any] = prepare_img() __magic_name__ : Tuple = processor(images=__A , return_tensors="""pt""" ).to(__A ) with torch.no_grad(): __magic_name__ : Union[str, Any] = model(**__A ) __magic_name__ : Union[str, Any] = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , __A ) __magic_name__ : Union[str, Any] = torch.tensor( [[-7.5_9_5_8, -7.5_9_5_8, -7.4_3_0_2], [-7.5_9_5_8, -7.5_9_5_8, -7.4_3_0_2], [-7.4_7_9_7, -7.4_7_9_7, -7.3_0_6_8]] ).to(__A ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , __A , atol=1e-4 ) ) def __magic_name__ ( self ) -> int: __magic_name__ : List[str] = AutoImageProcessor.from_pretrained("""openmmlab/upernet-convnext-tiny""" ) __magic_name__ : List[Any] = UperNetForSemanticSegmentation.from_pretrained("""openmmlab/upernet-convnext-tiny""" ).to(__A ) __magic_name__ : Any = prepare_img() __magic_name__ : Dict = processor(images=__A , return_tensors="""pt""" ).to(__A ) with torch.no_grad(): __magic_name__ : Tuple = model(**__A ) __magic_name__ : Dict = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , __A ) __magic_name__ : Dict = torch.tensor( [[-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.7_7_4_6, -8.7_7_4_6, -8.6_1_3_0]] ).to(__A ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , __A , atol=1e-4 ) )
361
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class snake_case__ ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None ) -> int: super().__init__() __magic_name__ : Any = pad_token_id __magic_name__ : Any = max_length __magic_name__ : List[str] = vocab __magic_name__ : List[Any] = merges __magic_name__ : int = BytePairTokenizer(lowerCAmelCase__ , lowerCAmelCase__ , sequence_length=lowerCAmelCase__ ) @classmethod def __magic_name__ ( cls , lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Any: __magic_name__ : Union[str, Any] = [""" """.join(lowerCAmelCase__ ) for m in tokenizer.bpe_ranks.keys()] __magic_name__ : Union[str, Any] = tokenizer.get_vocab() return cls(lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __magic_name__ ( cls , lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: __magic_name__ : Optional[Any] = GPTaTokenizer.from_pretrained(lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) return cls.from_tokenizer(lowerCAmelCase__ , *lowerCAmelCase__ , **lowerCAmelCase__ ) @classmethod def __magic_name__ ( cls , lowerCAmelCase__ ) -> List[Any]: return cls(**lowerCAmelCase__ ) def __magic_name__ ( self ) -> int: return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> int: __magic_name__ : Dict = self.tf_tokenizer(lowerCAmelCase__ ) __magic_name__ : Dict = tf.ones_like(lowerCAmelCase__ ) if self.pad_token_id is not None: # pad the tokens up to max length __magic_name__ : List[Any] = max_length if max_length is not None else self.max_length if max_length is not None: __magic_name__ ,__magic_name__ : List[Any] = pad_model_inputs( lowerCAmelCase__ , max_seq_length=lowerCAmelCase__ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
138
0
'''simple docstring''' import math def lowerCamelCase (_SCREAMING_SNAKE_CASE : int ): __a : int = math.loga(math.sqrt(4 * positive_integer + 1 ) / 2 + 1 / 2 ) return exponent == int(_SCREAMING_SNAKE_CASE ) def lowerCamelCase (_SCREAMING_SNAKE_CASE : float = 1 / 12_345 ): __a : List[str] = 0 __a : Any = 0 __a : int = 3 while True: __a : Tuple = (integer**2 - 1) / 4 # if candidate is an integer, then there is a partition for k if partition_candidate == int(_SCREAMING_SNAKE_CASE ): __a : str = int(_SCREAMING_SNAKE_CASE ) total_partitions += 1 if check_partition_perfect(_SCREAMING_SNAKE_CASE ): perfect_partitions += 1 if perfect_partitions > 0: if perfect_partitions / total_partitions < max_proportion: return int(_SCREAMING_SNAKE_CASE ) integer += 1 if __name__ == "__main__": print(f'''{solution() = }''')
27
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
0
"""simple docstring""" import math def lowerCamelCase_( _lowerCamelCase ) -> bool: '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_lowerCamelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCamelCase_( _lowerCamelCase = 10001 ) -> int: '''simple docstring''' try: _lowerCamelCase : Dict = int(_lowerCamelCase ) except (TypeError, ValueError): raise TypeError("Parameter nth must be int or castable to int." ) from None if nth <= 0: raise ValueError("Parameter nth must be greater than or equal to one." ) _lowerCamelCase : list[int] = [] _lowerCamelCase : int = 2 while len(_lowerCamelCase ) < nth: if is_prime(_lowerCamelCase ): primes.append(_lowerCamelCase ) num += 1 else: num += 1 return primes[len(_lowerCamelCase ) - 1] if __name__ == "__main__": print(f'''{solution() = }''')
340
"""simple docstring""" from datetime import datetime import matplotlib.pyplot as plt import torch def lowerCamelCase_( _lowerCamelCase ) -> Any: '''simple docstring''' for param in module.parameters(): _lowerCamelCase : Optional[int] = False def lowerCamelCase_( ) -> List[str]: '''simple docstring''' _lowerCamelCase : Optional[Any] = "cuda" if torch.cuda.is_available() else "cpu" if torch.backends.mps.is_available() and torch.backends.mps.is_built(): _lowerCamelCase : int = "mps" if device == "mps": print( "WARNING: MPS currently doesn't seem to work, and messes up backpropagation without any visible torch" " errors. I recommend using CUDA on a colab notebook or CPU instead if you're facing inexplicable issues" " with generations." ) return device def lowerCamelCase_( _lowerCamelCase ) -> Optional[int]: '''simple docstring''' _lowerCamelCase : Dict = plt.imshow(_lowerCamelCase ) fig.axes.get_xaxis().set_visible(_lowerCamelCase ) fig.axes.get_yaxis().set_visible(_lowerCamelCase ) plt.show() def lowerCamelCase_( ) -> str: '''simple docstring''' _lowerCamelCase : Tuple = datetime.now() _lowerCamelCase : Tuple = current_time.strftime("%H:%M:%S" ) return timestamp
340
1
from math import factorial def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): if successes > trials: raise ValueError("""successes must be lower or equal to trials""" ) if trials < 0 or successes < 0: raise ValueError("""the function is defined for non-negative integers""" ) if not isinstance(snake_case_,snake_case_ ) or not isinstance(snake_case_,snake_case_ ): raise ValueError("""the function is defined for non-negative integers""" ) if not 0 < prob < 1: raise ValueError("""prob has to be in range of 1 - 0""" ) _A : Union[str, Any] = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! _A : Any = float(factorial(snake_case_ ) ) coefficient /= factorial(snake_case_ ) * factorial(trials - successes ) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print("Probability of 2 successes out of 4 trails") print("with probability of 0.75 is:", end=" ") print(binomial_distribution(2, 4, 0.7_5))
26
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowercase_ = logging.get_logger(__name__) lowercase_ = { "microsoft/resnet-50": "https://huggingface.co/microsoft/resnet-50/blob/main/config.json", } class __A ( A , A ): '''simple docstring''' __lowerCamelCase : Tuple = 'resnet' __lowerCamelCase : Any = ['basic', 'bottleneck'] def __init__(self , A=3 , A=64 , A=[256, 512, 1_024, 2_048] , A=[3, 4, 6, 3] , A="bottleneck" , A="relu" , A=False , A=None , A=None , **A , ) -> Dict: """simple docstring""" super().__init__(**A ) if layer_type not in self.layer_types: raise ValueError(f'''layer_type={layer_type} is not one of {','.join(self.layer_types )}''' ) _a = num_channels _a = embedding_size _a = hidden_sizes _a = depths _a = layer_type _a = hidden_act _a = downsample_in_first_stage _a = ['''stem'''] + [f'''stage{idx}''' for idx in range(1 , len(A ) + 1 )] _a , _a = get_aligned_output_features_output_indices( out_features=A , out_indices=A , stage_names=self.stage_names ) class __A ( A ): '''simple docstring''' __lowerCamelCase : Any = version.parse('1.11' ) @property def a__ (self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def a__ (self ) -> float: """simple docstring""" return 1E-3
211
0
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase = 8 ) -> str: '''simple docstring''' lowerCAmelCase : Optional[int] = ascii_letters + digits + punctuation return "".join(secrets.choice(lowercase__ ) for _ in range(lowercase__ ) ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> str: '''simple docstring''' i -= len(lowercase__ ) lowerCAmelCase : Optional[int] = i // 3 lowerCAmelCase : Tuple = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) lowerCAmelCase : str = ( chars_incl + random(lowercase__, quotient + remainder ) + random(lowercase__, lowercase__ ) + random(lowercase__, lowercase__ ) ) lowerCAmelCase : Any = list(lowercase__ ) shuffle(lowercase__ ) return "".join(lowercase__ ) # random is a generalised function for letters, characters and numbers def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> str: '''simple docstring''' return "".join(secrets.choice(lowercase__ ) for _ in range(lowercase__ ) ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' pass # Put your code here... def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> Tuple: '''simple docstring''' pass # Put your code here... def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' pass # Put your code here... def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase = 8 ) -> bool: '''simple docstring''' if len(lowercase__ ) < min_length: # Your Password must be at least 8 characters long return False lowerCAmelCase : Tuple = any(char in ascii_uppercase for char in password ) lowerCAmelCase : Optional[Any] = any(char in ascii_lowercase for char in password ) lowerCAmelCase : Tuple = any(char in digits for char in password ) lowerCAmelCase : Optional[Any] = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def SCREAMING_SNAKE_CASE__ ( ) -> str: '''simple docstring''' lowerCAmelCase : int = int(input('Please indicate the max length of your password: ' ).strip() ) lowerCAmelCase : int = input( 'Please indicate the characters that must be in your password: ' ).strip() print('Password generated:', password_generator(lowercase__ ) ) print( 'Alternative Password generated:', alternative_password_generator(lowercase__, lowercase__ ), ) print('[If you are thinking of using this passsword, You better save it.]' ) if __name__ == "__main__": main()
356
from collections.abc import Sequence def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> float: '''simple docstring''' return sum(c * (x**i) for i, c in enumerate(_UpperCAmelCase ) ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> float: '''simple docstring''' lowerCAmelCase : Optional[int] = 0.0 for coeff in reversed(_UpperCAmelCase ): lowerCAmelCase : Union[str, Any] = result * x + coeff return result if __name__ == "__main__": __A : Optional[int] = (0.0, 0.0, 5.0, 9.3, 7.0) __A : str = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
323
0
'''simple docstring''' import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowercase : str = logging.get_logger(__name__) _lowercase : str = { "vocab_file": "vocab.json", "merges_file": "merges.txt", } _lowercase : Dict = { "vocab_file": {"ctrl": "https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-vocab.json"}, "merges_file": {"ctrl": "https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-merges.txt"}, } _lowercase : Union[str, Any] = { "ctrl": 2_5_6, } _lowercase : List[Any] = { "Pregnancy": 1_6_8_6_2_9, "Christianity": 7_6_7_5, "Explain": 1_0_6_4_2_3, "Fitness": 6_3_4_4_0, "Saving": 6_3_1_6_3, "Ask": 2_7_1_7_1, "Ass": 9_5_9_8_5, "Joke": 1_6_3_5_0_9, "Questions": 4_5_6_2_2, "Thoughts": 4_9_6_0_5, "Retail": 5_2_3_4_2, "Feminism": 1_6_4_3_3_8, "Writing": 1_1_9_9_2, "Atheism": 1_9_2_2_6_3, "Netflix": 4_8_6_1_6, "Computing": 3_9_6_3_9, "Opinion": 4_3_2_1_3, "Alone": 4_4_9_6_7, "Funny": 5_8_9_1_7, "Gaming": 4_0_3_5_8, "Human": 4_0_8_8, "India": 1_3_3_1, "Joker": 7_7_1_3_8, "Diet": 3_6_2_0_6, "Legal": 1_1_8_5_9, "Norman": 4_9_3_9, "Tip": 7_2_6_8_9, "Weight": 5_2_3_4_3, "Movies": 4_6_2_7_3, "Running": 2_3_4_2_5, "Science": 2_0_9_0, "Horror": 3_7_7_9_3, "Confession": 6_0_5_7_2, "Finance": 1_2_2_5_0, "Politics": 1_6_3_6_0, "Scary": 1_9_1_9_8_5, "Support": 1_2_6_5_4, "Technologies": 3_2_5_1_6, "Teenage": 6_6_1_6_0, "Event": 3_2_7_6_9, "Learned": 6_7_4_6_0, "Notion": 1_8_2_7_7_0, "Wikipedia": 3_7_5_8_3, "Books": 6_6_6_5, "Extract": 7_6_0_5_0, "Confessions": 1_0_2_7_0_1, "Conspiracy": 7_5_9_3_2, "Links": 6_3_6_7_4, "Narcissus": 1_5_0_4_2_5, "Relationship": 5_4_7_6_6, "Relationships": 1_3_4_7_9_6, "Reviews": 4_1_6_7_1, "News": 4_2_5_6, "Translation": 2_6_8_2_0, "multilingual": 1_2_8_4_0_6, } def snake_case_ ( __SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" lowercase_ : List[str] = set() lowercase_ : Tuple = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase_ : List[Any] = char lowercase_ : Union[str, Any] = set(__SCREAMING_SNAKE_CASE ) return pairs class lowerCAmelCase__ ( lowerCamelCase_ ): lowerCAmelCase_ = VOCAB_FILES_NAMES lowerCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ = CONTROL_CODES def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE="<unk>" , **__SCREAMING_SNAKE_CASE ): """simple docstring""" super().__init__(unk_token=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) with open(__SCREAMING_SNAKE_CASE , encoding='''utf-8''' ) as vocab_handle: lowercase_ : int = json.load(__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = {v: k for k, v in self.encoder.items()} with open(__SCREAMING_SNAKE_CASE , encoding='''utf-8''' ) as merges_handle: lowercase_ : Optional[Any] = merges_handle.read().split('''\n''' )[1:-1] lowercase_ : Tuple = [tuple(merge.split() ) for merge in merges] lowercase_ : Union[str, Any] = dict(zip(__SCREAMING_SNAKE_CASE , range(len(__SCREAMING_SNAKE_CASE ) ) ) ) lowercase_ : str = {} @property def _snake_case ( self ): """simple docstring""" return len(self.encoder ) def _snake_case ( self ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" if token in self.cache: return self.cache[token] lowercase_ : Tuple = tuple(__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) lowercase_ : Any = get_pairs(__SCREAMING_SNAKE_CASE ) if not pairs: return token while True: lowercase_ : Optional[Any] = min(__SCREAMING_SNAKE_CASE , key=lambda __SCREAMING_SNAKE_CASE : self.bpe_ranks.get(__SCREAMING_SNAKE_CASE , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break lowercase_ , lowercase_ : str = bigram lowercase_ : List[str] = [] lowercase_ : Tuple = 0 while i < len(__SCREAMING_SNAKE_CASE ): try: lowercase_ : int = word.index(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase_ : int = j if word[i] == first and i < len(__SCREAMING_SNAKE_CASE ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase_ : List[Any] = tuple(__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = new_word if len(__SCREAMING_SNAKE_CASE ) == 1: break else: lowercase_ : Tuple = get_pairs(__SCREAMING_SNAKE_CASE ) lowercase_ : Dict = '''@@ '''.join(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = word[:-4] lowercase_ : Union[str, Any] = word return word def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Union[str, Any] = [] lowercase_ : str = re.findall(R'''\S+\n?''' , __SCREAMING_SNAKE_CASE ) for token in words: split_tokens.extend(list(self.bpe(__SCREAMING_SNAKE_CASE ).split(''' ''' ) ) ) return split_tokens def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" return self.encoder.get(__SCREAMING_SNAKE_CASE , self.encoder.get(self.unk_token ) ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" return self.decoder.get(__SCREAMING_SNAKE_CASE , self.unk_token ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Tuple = ''' '''.join(__SCREAMING_SNAKE_CASE ).replace('''@@ ''' , '''''' ).strip() return out_string def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase_ : Tuple = os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase_ : int = os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(__SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__SCREAMING_SNAKE_CASE , ensure_ascii=__SCREAMING_SNAKE_CASE ) + '''\n''' ) lowercase_ : Dict = 0 with open(__SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __SCREAMING_SNAKE_CASE : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' ''' Please check that the tokenizer is not corrupted!''' ) lowercase_ : str = token_index writer.write(''' '''.join(__SCREAMING_SNAKE_CASE ) + '''\n''' ) index += 1 return vocab_file, merge_file # def decode(self, token_ids, skip_special_tokens=False, clean_up_tokenization_spaces=True): # filtered_tokens = ' '.join(self.convert_ids_to_tokens(token_ids, skip_special_tokens=skip_special_tokens)) # tokens_generated_so_far = re.sub('(@@ )', '', string=filtered_tokens) # tokens_generated_so_far = re.sub('(@@ ?$)', '', string=tokens_generated_so_far) # return ''.join(tokens_generated_so_far)
93
def lowercase__ ( __snake_case : Dict ): '''simple docstring''' if not head: return True # split the list to two parts UpperCAmelCase_ , UpperCAmelCase_ : Any = head.next, head while fast and fast.next: UpperCAmelCase_ : str = fast.next.next UpperCAmelCase_ : Union[str, Any] = slow.next UpperCAmelCase_ : int = slow.next UpperCAmelCase_ : List[Any] = None # Don't forget here! But forget still works! # reverse the second part UpperCAmelCase_ : Tuple = None while second: UpperCAmelCase_ : int = second.next UpperCAmelCase_ : Any = node UpperCAmelCase_ : Optional[Any] = second UpperCAmelCase_ : Tuple = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False UpperCAmelCase_ : Optional[Any] = node.next UpperCAmelCase_ : Dict = head.next return True def lowercase__ ( __snake_case : Union[str, Any] ): '''simple docstring''' if not head or not head.next: return True # 1. Get the midpoint (slow) UpperCAmelCase_ : Any = head while fast and fast.next: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = fast.next.next, slow.next # 2. Push the second half into the stack UpperCAmelCase_ : List[str] = [slow.val] while slow.next: UpperCAmelCase_ : List[str] = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False UpperCAmelCase_ : int = cur.next return True def lowercase__ ( __snake_case : Dict ): '''simple docstring''' if not head or not head.next: return True UpperCAmelCase_ : Tuple = {} UpperCAmelCase_ : int = 0 while head: if head.val in d: d[head.val].append(__snake_case ) else: UpperCAmelCase_ : List[Any] = [pos] UpperCAmelCase_ : Any = head.next pos += 1 UpperCAmelCase_ : Dict = pos - 1 UpperCAmelCase_ : Optional[int] = 0 for v in d.values(): if len(__snake_case ) % 2 != 0: middle += 1 else: UpperCAmelCase_ : int = 0 for i in range(0 , len(__snake_case ) ): if v[i] + v[len(__snake_case ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
29
0
'''simple docstring''' def a_ ( _UpperCAmelCase : float ) -> float: if edge <= 0 or not isinstance(_UpperCAmelCase ,_UpperCAmelCase ): raise ValueError('Length must be a positive.' ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def a_ ( _UpperCAmelCase : float ) -> float: if edge <= 0 or not isinstance(_UpperCAmelCase ,_UpperCAmelCase ): raise ValueError('Length must be a positive.' ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
364
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available A__ : int = { '''configuration_groupvit''': [ '''GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GroupViTConfig''', '''GroupViTOnnxConfig''', '''GroupViTTextConfig''', '''GroupViTVisionConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Tuple = [ '''GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GroupViTModel''', '''GroupViTPreTrainedModel''', '''GroupViTTextModel''', '''GroupViTVisionModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Optional[int] = [ '''TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFGroupViTModel''', '''TFGroupViTPreTrainedModel''', '''TFGroupViTTextModel''', '''TFGroupViTVisionModel''', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys A__ : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
0
0
"""simple docstring""" import numpy as np import qiskit def _snake_case ( lowercase__ = 8 , lowercase__ = None ): _lowerCamelCase : str = np.random.default_rng(seed=lowercase__ ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. _lowerCamelCase : List[str] = 6 * key_len # Measurement basis for Alice's qubits. _lowerCamelCase : int = rng.integers(2 , size=lowercase__ ) # The set of states Alice will prepare. _lowerCamelCase : str = rng.integers(2 , size=lowercase__ ) # Measurement basis for Bob's qubits. _lowerCamelCase : str = rng.integers(2 , size=lowercase__ ) # Quantum Circuit to simulate BB84 _lowerCamelCase : Dict = qiskit.QuantumCircuit(lowercase__ , name='BB84' ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(lowercase__ ): if alice_state[index] == 1: bbaa_circ.x(lowercase__ ) if alice_basis[index] == 1: bbaa_circ.h(lowercase__ ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(lowercase__ ): if bob_basis[index] == 1: bbaa_circ.h(lowercase__ ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. _lowerCamelCase : List[str] = qiskit.Aer.get_backend('aer_simulator' ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. _lowerCamelCase : List[Any] = qiskit.execute(lowercase__ , lowercase__ , shots=1 , seed_simulator=lowercase__ ) # Returns the result of measurement. _lowerCamelCase : Optional[Any] = job.result().get_counts(lowercase__ ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. _lowerCamelCase : Optional[int] = ''.join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( lowercase__ , lowercase__ , lowercase__ ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. _lowerCamelCase : Union[str, Any] = gen_key[:key_len] if len(lowercase__ ) >= key_len else gen_key.ljust(lowercase__ , '0' ) return key if __name__ == "__main__": print(F"The generated key is : {bbaa(8, seed=0)}") from doctest import testmod testmod()
96
import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated __A : Tuple = collections.namedtuple('''_Datasets''', ['''train''', '''validation''', '''test''']) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ __A : Tuple = '''https://storage.googleapis.com/cvdf-datasets/mnist/''' def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> Dict: '''simple docstring''' lowerCAmelCase : Dict = numpy.dtype(numpy.uintaa ).newbyteorder('>' ) return numpy.frombuffer(bytestream.read(4 ), dtype=_UpperCAmelCase )[0] @deprecated(_UpperCAmelCase, 'Please use tf.data to implement this functionality.' ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> int: '''simple docstring''' print('Extracting', f.name ) with gzip.GzipFile(fileobj=_UpperCAmelCase ) as bytestream: lowerCAmelCase : List[str] = _readaa(_UpperCAmelCase ) if magic != 2_051: raise ValueError( 'Invalid magic number %d in MNIST image file: %s' % (magic, f.name) ) lowerCAmelCase : Optional[Any] = _readaa(_UpperCAmelCase ) lowerCAmelCase : Any = _readaa(_UpperCAmelCase ) lowerCAmelCase : List[Any] = _readaa(_UpperCAmelCase ) lowerCAmelCase : Union[str, Any] = bytestream.read(rows * cols * num_images ) lowerCAmelCase : Any = numpy.frombuffer(_UpperCAmelCase, dtype=numpy.uinta ) lowerCAmelCase : Optional[int] = data.reshape(_UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, 1 ) return data @deprecated(_UpperCAmelCase, 'Please use tf.one_hot on tensors.' ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase ) -> Any: '''simple docstring''' lowerCAmelCase : Optional[Any] = labels_dense.shape[0] lowerCAmelCase : Union[str, Any] = numpy.arange(_UpperCAmelCase ) * num_classes lowerCAmelCase : List[str] = numpy.zeros((num_labels, num_classes) ) lowerCAmelCase : List[str] = 1 return labels_one_hot @deprecated(_UpperCAmelCase, 'Please use tf.data to implement this functionality.' ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase=False, _UpperCAmelCase=10 ) -> List[str]: '''simple docstring''' print('Extracting', f.name ) with gzip.GzipFile(fileobj=_UpperCAmelCase ) as bytestream: lowerCAmelCase : List[str] = _readaa(_UpperCAmelCase ) if magic != 2_049: raise ValueError( 'Invalid magic number %d in MNIST label file: %s' % (magic, f.name) ) lowerCAmelCase : Optional[Any] = _readaa(_UpperCAmelCase ) lowerCAmelCase : Dict = bytestream.read(_UpperCAmelCase ) lowerCAmelCase : Dict = numpy.frombuffer(_UpperCAmelCase, dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(_UpperCAmelCase, _UpperCAmelCase ) return labels class __A : @deprecated( UpperCAmelCase_ , 'Please use alternatives such as official/mnist/_DataSet.py' ' from tensorflow/models.' , ) def __init__( self : Any , UpperCAmelCase_ : int , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : str=False , UpperCAmelCase_ : int=False , UpperCAmelCase_ : List[str]=dtypes.floataa , UpperCAmelCase_ : Union[str, Any]=True , UpperCAmelCase_ : Optional[Any]=None , ): lowerCAmelCase , lowerCAmelCase : int = random_seed.get_seed(UpperCAmelCase_ ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowerCAmelCase : List[str] = dtypes.as_dtype(UpperCAmelCase_ ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError('Invalid image dtype %r, expected uint8 or float32' % dtype ) if fake_data: lowerCAmelCase : Dict = 10000 lowerCAmelCase : Any = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f"images.shape: {images.shape} labels.shape: {labels.shape}" lowerCAmelCase : Optional[Any] = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowerCAmelCase : Union[str, Any] = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowerCAmelCase : Optional[int] = images.astype(numpy.floataa ) lowerCAmelCase : Dict = numpy.multiply(UpperCAmelCase_ , 1.0 / 2_55.0 ) lowerCAmelCase : List[str] = images lowerCAmelCase : List[str] = labels lowerCAmelCase : List[Any] = 0 lowerCAmelCase : Optional[int] = 0 @property def lowercase__ ( self : str ): return self._images @property def lowercase__ ( self : Dict ): return self._labels @property def lowercase__ ( self : List[Any] ): return self._num_examples @property def lowercase__ ( self : Any ): return self._epochs_completed def lowercase__ ( self : Optional[Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[int]=False , UpperCAmelCase_ : List[str]=True ): if fake_data: lowerCAmelCase : Union[str, Any] = [1] * 784 lowerCAmelCase : Dict = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(UpperCAmelCase_ )], [fake_label for _ in range(UpperCAmelCase_ )], ) lowerCAmelCase : Union[str, Any] = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowerCAmelCase : Optional[int] = numpy.arange(self._num_examples ) numpy.random.shuffle(UpperCAmelCase_ ) lowerCAmelCase : List[Any] = self.images[perma] lowerCAmelCase : str = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowerCAmelCase : Tuple = self._num_examples - start lowerCAmelCase : Union[str, Any] = self._images[start : self._num_examples] lowerCAmelCase : Tuple = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowerCAmelCase : Dict = numpy.arange(self._num_examples ) numpy.random.shuffle(UpperCAmelCase_ ) lowerCAmelCase : List[Any] = self.images[perm] lowerCAmelCase : Optional[Any] = self.labels[perm] # Start next epoch lowerCAmelCase : Optional[Any] = 0 lowerCAmelCase : Dict = batch_size - rest_num_examples lowerCAmelCase : int = self._index_in_epoch lowerCAmelCase : Union[str, Any] = self._images[start:end] lowerCAmelCase : int = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowerCAmelCase : Optional[Any] = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(_UpperCAmelCase, 'Please write your own downloading logic.' ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ) -> Any: '''simple docstring''' if not gfile.Exists(_UpperCAmelCase ): gfile.MakeDirs(_UpperCAmelCase ) lowerCAmelCase : Union[str, Any] = os.path.join(_UpperCAmelCase, _UpperCAmelCase ) if not gfile.Exists(_UpperCAmelCase ): urllib.request.urlretrieve(_UpperCAmelCase, _UpperCAmelCase ) # noqa: S310 with gfile.GFile(_UpperCAmelCase ) as f: lowerCAmelCase : List[Any] = f.size() print('Successfully downloaded', _UpperCAmelCase, _UpperCAmelCase, 'bytes.' ) return filepath @deprecated( _UpperCAmelCase, 'Please use alternatives such as:' ' tensorflow_datasets.load(\'mnist\')' ) def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase, _UpperCAmelCase=False, _UpperCAmelCase=False, _UpperCAmelCase=dtypes.floataa, _UpperCAmelCase=True, _UpperCAmelCase=5_000, _UpperCAmelCase=None, _UpperCAmelCase=DEFAULT_SOURCE_URL, ) -> Tuple: '''simple docstring''' if fake_data: def fake(): return _DataSet( [], [], fake_data=_UpperCAmelCase, one_hot=_UpperCAmelCase, dtype=_UpperCAmelCase, seed=_UpperCAmelCase ) lowerCAmelCase : Tuple = fake() lowerCAmelCase : Optional[Any] = fake() lowerCAmelCase : List[Any] = fake() return _Datasets(train=_UpperCAmelCase, validation=_UpperCAmelCase, test=_UpperCAmelCase ) if not source_url: # empty string check lowerCAmelCase : Any = DEFAULT_SOURCE_URL lowerCAmelCase : Optional[Any] = 'train-images-idx3-ubyte.gz' lowerCAmelCase : Any = 'train-labels-idx1-ubyte.gz' lowerCAmelCase : int = 't10k-images-idx3-ubyte.gz' lowerCAmelCase : Union[str, Any] = 't10k-labels-idx1-ubyte.gz' lowerCAmelCase : str = _maybe_download( _UpperCAmelCase, _UpperCAmelCase, source_url + train_images_file ) with gfile.Open(_UpperCAmelCase, 'rb' ) as f: lowerCAmelCase : Any = _extract_images(_UpperCAmelCase ) lowerCAmelCase : Tuple = _maybe_download( _UpperCAmelCase, _UpperCAmelCase, source_url + train_labels_file ) with gfile.Open(_UpperCAmelCase, 'rb' ) as f: lowerCAmelCase : int = _extract_labels(_UpperCAmelCase, one_hot=_UpperCAmelCase ) lowerCAmelCase : Optional[Any] = _maybe_download( _UpperCAmelCase, _UpperCAmelCase, source_url + test_images_file ) with gfile.Open(_UpperCAmelCase, 'rb' ) as f: lowerCAmelCase : List[Any] = _extract_images(_UpperCAmelCase ) lowerCAmelCase : Any = _maybe_download( _UpperCAmelCase, _UpperCAmelCase, source_url + test_labels_file ) with gfile.Open(_UpperCAmelCase, 'rb' ) as f: lowerCAmelCase : List[str] = _extract_labels(_UpperCAmelCase, one_hot=_UpperCAmelCase ) if not 0 <= validation_size <= len(_UpperCAmelCase ): lowerCAmelCase : str = ( 'Validation size should be between 0 and ' f"{len(_UpperCAmelCase )}. Received: {validation_size}." ) raise ValueError(_UpperCAmelCase ) lowerCAmelCase : str = train_images[:validation_size] lowerCAmelCase : Dict = train_labels[:validation_size] lowerCAmelCase : List[str] = train_images[validation_size:] lowerCAmelCase : str = train_labels[validation_size:] lowerCAmelCase : str = {'dtype': dtype, 'reshape': reshape, 'seed': seed} lowerCAmelCase : int = _DataSet(_UpperCAmelCase, _UpperCAmelCase, **_UpperCAmelCase ) lowerCAmelCase : Union[str, Any] = _DataSet(_UpperCAmelCase, _UpperCAmelCase, **_UpperCAmelCase ) lowerCAmelCase : Union[str, Any] = _DataSet(_UpperCAmelCase, _UpperCAmelCase, **_UpperCAmelCase ) return _Datasets(train=_UpperCAmelCase, validation=_UpperCAmelCase, test=_UpperCAmelCase )
138
0
from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = 'EncodecFeatureExtractor' __magic_name__ = ('T5Tokenizer', 'T5TokenizerFast') def __init__( self , __snake_case , __snake_case ): super().__init__(__snake_case , __snake_case ) snake_case = self.feature_extractor snake_case = False def a_ ( self , __snake_case=None , __snake_case=None , __snake_case=True ): return self.tokenizer.get_decoder_prompt_ids(task=__snake_case , language=__snake_case , no_timestamps=__snake_case ) def __call__( self , *__snake_case , **__snake_case ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*__snake_case , **__snake_case ) snake_case = kwargs.pop('''audio''' , __snake_case ) snake_case = kwargs.pop('''sampling_rate''' , __snake_case ) snake_case = kwargs.pop('''text''' , __snake_case ) if len(__snake_case ) > 0: snake_case = args[0] snake_case = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if text is not None: snake_case = self.tokenizer(__snake_case , **__snake_case ) if audio is not None: snake_case = self.feature_extractor(__snake_case , *__snake_case , sampling_rate=__snake_case , **__snake_case ) if audio is None: return inputs elif text is None: return audio_inputs else: snake_case = audio_inputs['''input_values'''] if "padding_mask" in audio_inputs: snake_case = audio_inputs['''padding_mask'''] return inputs def a_ ( self , *__snake_case , **__snake_case ): snake_case = kwargs.pop('''audio''' , __snake_case ) snake_case = kwargs.pop('''padding_mask''' , __snake_case ) if len(__snake_case ) > 0: snake_case = args[0] snake_case = args[1:] if audio_values is not None: return self._decode_audio(__snake_case , padding_mask=__snake_case ) else: return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def a_ ( self , *__snake_case , **__snake_case ): return self.tokenizer.decode(*__snake_case , **__snake_case ) def a_ ( self , __snake_case , __snake_case = None ): snake_case = to_numpy(__snake_case ) snake_case , snake_case , snake_case = audio_values.shape if padding_mask is None: return list(__snake_case ) snake_case = to_numpy(__snake_case ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) snake_case = seq_len - padding_mask.shape[-1] snake_case = 1 - self.feature_extractor.padding_value snake_case = np.pad(__snake_case , ((0, 0), (0, difference)) , '''constant''' , constant_values=__snake_case ) snake_case = audio_values.tolist() for i in range(__snake_case ): snake_case = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] snake_case = sliced_audio.reshape(__snake_case , -1 ) return audio_values
213
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class A__ : """simple docstring""" def __init__( self , __snake_case , __snake_case=1_3 , __snake_case=7 , __snake_case=True , __snake_case=True , __snake_case=True , __snake_case=True , __snake_case=9_9 , __snake_case=6_4 , __snake_case=5 , __snake_case=4 , __snake_case=3_7 , __snake_case="gelu" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=5_1_2 , __snake_case=1_6 , __snake_case=2 , __snake_case=0.02 , __snake_case=3 , __snake_case=4 , __snake_case=None , ): snake_case = parent snake_case = batch_size snake_case = seq_length snake_case = is_training snake_case = use_input_mask snake_case = use_token_type_ids snake_case = use_labels snake_case = vocab_size snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = max_position_embeddings snake_case = type_vocab_size snake_case = type_sequence_label_size snake_case = initializer_range snake_case = num_labels snake_case = num_choices snake_case = scope snake_case = vocab_size - 1 def a_ ( self ): snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case = None if self.use_input_mask: snake_case = random_attention_mask([self.batch_size, self.seq_length] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case = self.get_config() return config, input_ids, input_mask, token_labels def a_ ( self ): return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__snake_case , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def a_ ( self ): snake_case , snake_case , snake_case , snake_case = self.prepare_config_and_inputs() snake_case = True return config, input_ids, input_mask, token_labels def a_ ( self , __snake_case , __snake_case , __snake_case ): snake_case = GPTNeoXModel(config=__snake_case ) model.to(__snake_case ) model.eval() snake_case = model(__snake_case , attention_mask=__snake_case ) snake_case = model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a_ ( self , __snake_case , __snake_case , __snake_case ): snake_case = True snake_case = GPTNeoXModel(__snake_case ) model.to(__snake_case ) model.eval() snake_case = model(__snake_case , attention_mask=__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a_ ( self , __snake_case , __snake_case , __snake_case , __snake_case ): snake_case = GPTNeoXForCausalLM(config=__snake_case ) model.to(__snake_case ) model.eval() snake_case = model(__snake_case , attention_mask=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a_ ( self , __snake_case , __snake_case , __snake_case , __snake_case ): snake_case = self.num_labels snake_case = GPTNeoXForQuestionAnswering(__snake_case ) model.to(__snake_case ) model.eval() snake_case = model(__snake_case , attention_mask=__snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a_ ( self , __snake_case , __snake_case , __snake_case , __snake_case ): snake_case = self.num_labels snake_case = GPTNeoXForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = model(__snake_case , attention_mask=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a_ ( self , __snake_case , __snake_case , __snake_case , __snake_case ): snake_case = self.num_labels snake_case = GPTNeoXForTokenClassification(__snake_case ) model.to(__snake_case ) model.eval() snake_case = model(__snake_case , attention_mask=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a_ ( self , __snake_case , __snake_case , __snake_case ): snake_case = True snake_case = GPTNeoXForCausalLM(config=__snake_case ) model.to(__snake_case ) model.eval() # first forward pass snake_case = model(__snake_case , attention_mask=__snake_case , use_cache=__snake_case ) snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case = model(__snake_case , attention_mask=__snake_case , output_hidden_states=__snake_case ) snake_case = output_from_no_past['''hidden_states'''][0] snake_case = model( __snake_case , attention_mask=__snake_case , past_key_values=__snake_case , output_hidden_states=__snake_case , )['''hidden_states'''][0] # select random slice snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__snake_case , __snake_case , atol=1E-3 ) ) def a_ ( self ): snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class A__ ( snake_case__ , snake_case__ , snake_case__ , unittest.TestCase ): """simple docstring""" __magic_name__ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) __magic_name__ = (GPTNeoXForCausalLM,) if is_torch_available() else () __magic_name__ = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) __magic_name__ = False __magic_name__ = False __magic_name__ = False __magic_name__ = False def a_ ( self ): snake_case = GPTNeoXModelTester(self ) snake_case = ConfigTester(self , config_class=__snake_case , hidden_size=6_4 , num_attention_heads=8 ) def a_ ( self ): self.config_tester.run_common_tests() def a_ ( self ): snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(__snake_case , __snake_case , __snake_case ) def a_ ( self ): snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(__snake_case , __snake_case , __snake_case ) def a_ ( self ): # This regression test was failing with PyTorch < 1.3 snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_decoder() snake_case = None self.model_tester.create_and_check_model_as_decoder(__snake_case , __snake_case , __snake_case ) def a_ ( self ): snake_case , snake_case , snake_case , snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(__snake_case , __snake_case , __snake_case ) def a_ ( self ): snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*__snake_case ) def a_ ( self ): snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__snake_case ) def a_ ( self ): snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__snake_case ) def a_ ( self ): snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__snake_case ) @unittest.skip(reason='''Feed forward chunking is not implemented''' ) def a_ ( self ): pass @parameterized.expand([('''linear''',), ('''dynamic''',)] ) def a_ ( self , __snake_case ): snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = ids_tensor([1, 1_0] , config.vocab_size ) snake_case = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights snake_case = GPTNeoXModel(__snake_case ) original_model.to(__snake_case ) original_model.eval() snake_case = original_model(__snake_case ).last_hidden_state snake_case = original_model(__snake_case ).last_hidden_state set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights snake_case = {'''type''': scaling_type, '''factor''': 10.0} snake_case = GPTNeoXModel(__snake_case ) scaled_model.to(__snake_case ) scaled_model.eval() snake_case = scaled_model(__snake_case ).last_hidden_state snake_case = scaled_model(__snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(__snake_case , __snake_case , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(__snake_case , __snake_case , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(__snake_case , __snake_case , atol=1E-5 ) ) @require_torch class A__ ( unittest.TestCase ): """simple docstring""" @slow def a_ ( self ): snake_case = AutoTokenizer.from_pretrained('''EleutherAI/pythia-410m-deduped''' ) for checkpointing in [True, False]: snake_case = GPTNeoXForCausalLM.from_pretrained('''EleutherAI/pythia-410m-deduped''' ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(__snake_case ) snake_case = tokenizer('''My favorite food is''' , return_tensors='''pt''' ).to(__snake_case ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 snake_case = '''My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI\'m not sure''' snake_case = model.generate(**__snake_case , do_sample=__snake_case , max_new_tokens=2_0 ) snake_case = tokenizer.batch_decode(__snake_case )[0] self.assertEqual(__snake_case , __snake_case )
213
1
import math def _a ( UpperCamelCase_ : int ) -> bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(UpperCamelCase_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _a ( UpperCamelCase_ : int = 10_001 ) -> int: """simple docstring""" try: lowerCAmelCase__ = int(UpperCamelCase_ ) except (TypeError, ValueError): raise TypeError("Parameter nth must be int or castable to int." ) from None if nth <= 0: raise ValueError("Parameter nth must be greater than or equal to one." ) lowerCAmelCase__ = [] lowerCAmelCase__ = 2 while len(UpperCamelCase_ ) < nth: if is_prime(UpperCamelCase_ ): primes.append(UpperCamelCase_ ) num += 1 else: num += 1 return primes[len(UpperCamelCase_ ) - 1] if __name__ == "__main__": print(F"{solution() = }")
340
import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) a_ = {name: getattr(transformers, name + '''Fast''') for name in SLOW_TO_FAST_CONVERTERS} def _a ( UpperCamelCase_ : Any , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple , UpperCamelCase_ : Tuple ) -> List[str]: """simple docstring""" if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(F"Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}." ) if tokenizer_name is None: lowerCAmelCase__ = TOKENIZER_CLASSES else: lowerCAmelCase__ = {tokenizer_name: getattr(UpperCamelCase_ , tokenizer_name + "Fast" )} logger.info(F"Loading tokenizer classes: {tokenizer_names}" ) for tokenizer_name in tokenizer_names: lowerCAmelCase__ = TOKENIZER_CLASSES[tokenizer_name] lowerCAmelCase__ = True if checkpoint_name is None: lowerCAmelCase__ = list(tokenizer_class.max_model_input_sizes.keys() ) else: lowerCAmelCase__ = [checkpoint_name] logger.info(F"For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}" ) for checkpoint in checkpoint_names: logger.info(F"Loading {tokenizer_class.__class__.__name__} {checkpoint}" ) # Load tokenizer lowerCAmelCase__ = tokenizer_class.from_pretrained(UpperCamelCase_ , force_download=UpperCamelCase_ ) # Save fast tokenizer logger.info(F"Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}" ) # For organization names we create sub-directories if "/" in checkpoint: lowerCAmelCase__ , lowerCAmelCase__ = checkpoint.split("/" ) lowerCAmelCase__ = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) elif add_prefix: lowerCAmelCase__ = checkpoint lowerCAmelCase__ = dump_path else: lowerCAmelCase__ = None lowerCAmelCase__ = dump_path logger.info(F"=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}" ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: lowerCAmelCase__ = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] lowerCAmelCase__ = file_path.split(UpperCamelCase_ )[-1][0] if next_char == "/": lowerCAmelCase__ = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) lowerCAmelCase__ = None logger.info(F"=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}" ) lowerCAmelCase__ = tokenizer.save_pretrained( UpperCamelCase_ , legacy_format=UpperCamelCase_ , filename_prefix=UpperCamelCase_ ) logger.info(F"=> File names {file_names}" ) for file_name in file_names: if not file_name.endswith("tokenizer.json" ): os.remove(UpperCamelCase_ ) logger.info(F"=> removing {file_name}" ) if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output generated fast tokenizer files.''' ) parser.add_argument( '''--tokenizer_name''', default=None, type=str, help=( F"Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will " '''download and convert all the checkpoints from AWS.''' ), ) parser.add_argument( '''--checkpoint_name''', default=None, type=str, help='''Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.''', ) parser.add_argument( '''--force_download''', action='''store_true''', help='''Re-download checkpoints.''', ) a_ = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
340
1
"""simple docstring""" import operator as op a : List[Any] = '''scaler.pt''' a : Any = '''pytorch_model''' a : Union[str, Any] = '''random_states''' a : List[Any] = '''optimizer''' a : List[str] = '''scheduler''' a : Dict = '''pytorch_model.bin''' a : List[str] = '''pytorch_model.bin.index.json''' a : Optional[int] = '''model.safetensors''' a : List[str] = '''model.safetensors.index.json''' a : Optional[int] = '''1.10.2''' a : Dict = '''py38''' a : Dict = '''4.17.0''' a : Union[str, Any] = ['''ml.p3.16xlarge''', '''ml.p3dn.24xlarge''', '''ml.p4dn.24xlarge'''] a : int = ['''FULL_SHARD''', '''SHARD_GRAD_OP''', '''NO_SHARD''', '''HYBRID_SHARD''', '''HYBRID_SHARD_ZERO2'''] a : Tuple = ['''TRANSFORMER_BASED_WRAP''', '''SIZE_BASED_WRAP''', '''NO_WRAP'''] a : Tuple = ['''BACKWARD_PRE''', '''BACKWARD_POST''', '''NO_PREFETCH'''] a : List[Any] = ['''FULL_STATE_DICT''', '''LOCAL_STATE_DICT''', '''SHARDED_STATE_DICT'''] a : Dict = '''2.0.1''' a : Tuple = ['''pdsh''', '''standard''', '''openmpi''', '''mvapich'''] a : str = ['''default''', '''reduce-overhead''', '''max-autotune'''] a : Any = {'''>''': op.gt, '''>=''': op.ge, '''==''': op.eq, '''!=''': op.ne, '''<=''': op.le, '''<''': op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 a : List[Any] = [ '''nnodes''', '''nproc_per_node''', '''rdzv_backend''', '''rdzv_endpoint''', '''rdzv_id''', '''rdzv_conf''', '''standalone''', '''max_restarts''', '''monitor_interval''', '''start_method''', '''role''', '''module''', '''m''', '''no_python''', '''run_path''', '''log_dir''', '''r''', '''redirects''', '''t''', '''tee''', '''node_rank''', '''master_addr''', '''master_port''', ] a : str = ['''DEEPSPEED''', '''MULTI_GPU''', '''FSDP''', '''MEGATRON_LM'''] a : Dict = ['''DEEPSPEED''', '''MULTI_XPU''', '''FSDP''']
79
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->bool: '''simple docstring''' return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
79
1
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters _lowerCamelCase : int = (720, 1280) # Height, Width _lowerCamelCase : List[str] = (0.4, 0.6) # if height or width lower than this scale, drop it. _lowerCamelCase : int = 1 / 100 _lowerCamelCase : Optional[Any] = "" _lowerCamelCase : str = "" _lowerCamelCase : str = "" _lowerCamelCase : str = 250 def __lowerCamelCase ( ) -> None: """simple docstring""" UpperCamelCase , UpperCamelCase = get_dataset(A__ , A__ ) for index in range(A__ ): UpperCamelCase = random.sample(range(len(A__ ) ) , 4 ) UpperCamelCase , UpperCamelCase , UpperCamelCase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' UpperCamelCase = random_chars(32 ) UpperCamelCase = path.split(os.sep )[-1].rsplit('.' , 1 )[0] UpperCamelCase = F"""{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}""" cva.imwrite(F"""{file_root}.jpg""" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"""Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}""" ) UpperCamelCase = [] for anno in new_annos: UpperCamelCase = anno[3] - anno[1] UpperCamelCase = anno[4] - anno[2] UpperCamelCase = anno[1] + width / 2 UpperCamelCase = anno[2] + height / 2 UpperCamelCase = F"""{anno[0]} {x_center} {y_center} {width} {height}""" annos_list.append(A__ ) with open(F"""{file_root}.txt""" , 'w' ) as outfile: outfile.write('\n'.join(line for line in annos_list ) ) def __lowerCamelCase ( A__ , A__ ) -> tuple[list, list]: """simple docstring""" UpperCamelCase = [] UpperCamelCase = [] for label_file in glob.glob(os.path.join(A__ , '*.txt' ) ): UpperCamelCase = label_file.split(os.sep )[-1].rsplit('.' , 1 )[0] with open(A__ ) as in_file: UpperCamelCase = in_file.readlines() UpperCamelCase = os.path.join(A__ , F"""{label_name}.jpg""" ) UpperCamelCase = [] for obj_list in obj_lists: UpperCamelCase = obj_list.rstrip('\n' ).split(' ' ) UpperCamelCase = float(obj[1] ) - float(obj[3] ) / 2 UpperCamelCase = float(obj[2] ) - float(obj[4] ) / 2 UpperCamelCase = float(obj[1] ) + float(obj[3] ) / 2 UpperCamelCase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def __lowerCamelCase ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ) -> tuple[list, list, str]: """simple docstring""" UpperCamelCase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) UpperCamelCase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) UpperCamelCase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) UpperCamelCase = int(scale_x * output_size[1] ) UpperCamelCase = int(scale_y * output_size[0] ) UpperCamelCase = [] UpperCamelCase = [] for i, index in enumerate(A__ ): UpperCamelCase = all_img_list[index] path_list.append(A__ ) UpperCamelCase = all_annos[index] UpperCamelCase = cva.imread(A__ ) if i == 0: # top-left UpperCamelCase = cva.resize(A__ , (divid_point_x, divid_point_y) ) UpperCamelCase = img for bbox in img_annos: UpperCamelCase = bbox[1] * scale_x UpperCamelCase = bbox[2] * scale_y UpperCamelCase = bbox[3] * scale_x UpperCamelCase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right UpperCamelCase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) UpperCamelCase = img for bbox in img_annos: UpperCamelCase = scale_x + bbox[1] * (1 - scale_x) UpperCamelCase = bbox[2] * scale_y UpperCamelCase = scale_x + bbox[3] * (1 - scale_x) UpperCamelCase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left UpperCamelCase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) UpperCamelCase = img for bbox in img_annos: UpperCamelCase = bbox[1] * scale_x UpperCamelCase = scale_y + bbox[2] * (1 - scale_y) UpperCamelCase = bbox[3] * scale_x UpperCamelCase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right UpperCamelCase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) UpperCamelCase = img for bbox in img_annos: UpperCamelCase = scale_x + bbox[1] * (1 - scale_x) UpperCamelCase = scale_y + bbox[2] * (1 - scale_y) UpperCamelCase = scale_x + bbox[3] * (1 - scale_x) UpperCamelCase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: UpperCamelCase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def __lowerCamelCase ( A__ ) -> str: """simple docstring""" assert number_char > 1, "The number of character should greater than 1" UpperCamelCase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print("DONE ✅")
28
'''simple docstring''' import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class UpperCamelCase__ ( lowercase_ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ = FlaxAutoencoderKL @property def lowerCamelCase_ ( self : List[Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = 4 SCREAMING_SNAKE_CASE : str = 3 SCREAMING_SNAKE_CASE : List[Any] = (32, 32) SCREAMING_SNAKE_CASE : Tuple = jax.random.PRNGKey(0 ) SCREAMING_SNAKE_CASE : Any = jax.random.uniform(lowerCamelCase_ , ((batch_size, num_channels) + sizes) ) return {"sample": image, "prng_key": prng_key} def lowerCamelCase_ ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = { """block_out_channels""": [32, 64], """in_channels""": 3, """out_channels""": 3, """down_block_types""": ["""DownEncoderBlock2D""", """DownEncoderBlock2D"""], """up_block_types""": ["""UpDecoderBlock2D""", """UpDecoderBlock2D"""], """latent_channels""": 4, } SCREAMING_SNAKE_CASE : List[Any] = self.dummy_input return init_dict, inputs_dict
323
0
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class lowerCamelCase (unittest.TestCase ): """simple docstring""" UpperCAmelCase_ = inspect.getfile(accelerate.test_utils ) UpperCAmelCase_ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_cli.py"] ) UpperCAmelCase_ = ["accelerate", "launch"] UpperCAmelCase_ = Path.home() / ".cache/huggingface/accelerate" UpperCAmelCase_ = "default_config.yaml" UpperCAmelCase_ = config_folder / config_file UpperCAmelCase_ = config_folder / "_default_config.yaml" UpperCAmelCase_ = Path("tests/test_configs" ) @classmethod def A_ ( cls : List[Any] ) -> List[Any]: """simple docstring""" if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def A_ ( cls : Optional[int] ) -> List[str]: """simple docstring""" if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def A_ ( self : Union[str, Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path], env=os.environ.copy() ) def A_ ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" for config in sorted(self.test_config_path.glob("**/*.yaml" ) ): with self.subTest(config_file=_UpperCAmelCase ): execute_subprocess_async( self.base_cmd + ["--config_file", str(_UpperCAmelCase ), self.test_file_path], env=os.environ.copy() ) def A_ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" execute_subprocess_async(["accelerate", "test"], env=os.environ.copy() ) class lowerCamelCase (unittest.TestCase ): """simple docstring""" UpperCAmelCase_ = "test-tpu" UpperCAmelCase_ = "us-central1-a" UpperCAmelCase_ = "ls" UpperCAmelCase_ = ["accelerate", "tpu-config"] UpperCAmelCase_ = "cd /usr/share" UpperCAmelCase_ = "tests/test_samples/test_command_file.sh" UpperCAmelCase_ = "Running gcloud compute tpus tpu-vm ssh" def A_ ( self : str ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = run_command( self.cmd + ["--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug"], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''', _UpperCAmelCase, ) def A_ ( self : Any ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''', _UpperCAmelCase, ) def A_ ( self : List[str] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--debug"], return_stdout=_UpperCAmelCase ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''', _UpperCAmelCase, ) def A_ ( self : List[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--debug"], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all''', _UpperCAmelCase, ) def A_ ( self : int ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--command", "echo \"Hello World\"", "--debug", ], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo "Hello World" --worker all''', _UpperCAmelCase, ) def A_ ( self : Dict ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command_file", self.command_file, "--debug"], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''', _UpperCAmelCase, ) def A_ ( self : Tuple ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command_file", self.command_file, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo "hello world"; echo "this is a second command" --worker all''', _UpperCAmelCase, ) def A_ ( self : int ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--debug"], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo "hello world"; echo "this is a second command" --worker all''', _UpperCAmelCase, ) def A_ ( self : List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--accelerate_version", "12.0.0", "--debug", ], return_stdout=_UpperCAmelCase, ) self.assertIn( F'''{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo "hello world"; echo "this is a second command" --worker all''', _UpperCAmelCase, )
359
import requests from bsa import BeautifulSoup def _a ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : dict ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE__ , params=SCREAMING_SNAKE_CASE__ ).content , "html.parser" ) SCREAMING_SNAKE_CASE__ : str = soup.find("div" , attrs={"class": "gs_ri"} ) SCREAMING_SNAKE_CASE__ : int = div.find("div" , attrs={"class": "gs_fl"} ).find_all("a" ) return anchors[2].get_text() if __name__ == "__main__": _lowerCamelCase : Any = { '''title''': ( '''Precisely geometry controlled microsupercapacitors for ultrahigh areal ''' '''capacitance, volumetric capacitance, and energy density''' ), '''journal''': '''Chem. Mater.''', '''volume''': 3_0, '''pages''': '''3979-3990''', '''year''': 2_0_1_8, '''hl''': '''en''', } print(get_citation('''https://scholar.google.com/scholar_lookup''', params=params))
191
0
def a ( A__ : int , A__ : bool = False ) -> bool: """simple docstring""" if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not allow_probable: raise ValueError( 'Warning: upper bound of deterministic test is exceeded. ' 'Pass allow_probable=True to allow probabilistic test. ' 'A return value of True indicates a probable prime.' ) # array bounds provided by analysis _lowercase =[ 2047, 1373653, 25326001, 3215031751, 2152302898747, 3474749660383, 341550071728321, 1, 3825123056546413051, 1, 1, 318665857834031151167461, 3317044064679887385961981, ] _lowercase =[2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(A__ , 1 ): if n < _p: # then we have our last prime to check _lowercase =primes[:idx] break _lowercase , _lowercase =n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: _lowercase =False for r in range(A__ ): _lowercase =pow(A__ , d * 2**r , A__ ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): _lowercase =True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def a ( ) -> None: """simple docstring""" assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838201 ) assert miller_rabin(838207 ) # 1_373_653 assert not miller_rabin(17316001 ) assert miller_rabin(17316017 ) # 25_326_001 assert not miller_rabin(3078386641 ) assert miller_rabin(3078386653 ) # 3_215_031_751 assert not miller_rabin(1713045574801 ) assert miller_rabin(1713045574819 ) # 2_152_302_898_747 assert not miller_rabin(2779799728307 ) assert miller_rabin(2779799728327 ) # 3_474_749_660_383 assert not miller_rabin(113850023909441 ) assert miller_rabin(113850023909527 ) # 341_550_071_728_321 assert not miller_rabin(1275041018848804351 ) assert miller_rabin(1275041018848804391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79666464458507787791867 ) assert miller_rabin(79666464458507787791951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552840677446647897660333 ) assert miller_rabin(552840677446647897660359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
205
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase__ = { "configuration_groupvit": [ "GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "GroupViTConfig", "GroupViTOnnxConfig", "GroupViTTextConfig", "GroupViTVisionConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ "GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "GroupViTModel", "GroupViTPreTrainedModel", "GroupViTTextModel", "GroupViTVisionModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ "TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFGroupViTModel", "TFGroupViTPreTrainedModel", "TFGroupViTTextModel", "TFGroupViTVisionModel", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
0
0
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
362
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: SCREAMING_SNAKE_CASE__:Optional[Any] = None SCREAMING_SNAKE_CASE__:Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:int = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Optional[int] = { """vocab_file""": { """moussaKam/mbarthez""": """https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez""": """https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez-orangesum-title""": ( """https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """moussaKam/mbarthez""": """https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json""", """moussaKam/barthez""": """https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json""", """moussaKam/barthez-orangesum-title""": ( """https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE__:Dict = { """moussaKam/mbarthez""": 1024, """moussaKam/barthez""": 1024, """moussaKam/barthez-orangesum-title""": 1024, } SCREAMING_SNAKE_CASE__:List[str] = """▁""" class snake_case__ ( snake_case_ ): _snake_case : str = VOCAB_FILES_NAMES _snake_case : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : Dict = ["""input_ids""", """attention_mask"""] _snake_case : str = BarthezTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="</s>" , lowerCamelCase="<s>" , lowerCamelCase="<unk>" , lowerCamelCase="<pad>" , lowerCamelCase="<mask>" , **lowerCamelCase , ): # Mask token behave like a normal word, i.e. include the space before it __a = AddedToken(lowerCamelCase , lstrip=lowerCamelCase , rstrip=lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ) else mask_token super().__init__( lowerCamelCase , tokenizer_file=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , unk_token=lowerCamelCase , sep_token=lowerCamelCase , cls_token=lowerCamelCase , pad_token=lowerCamelCase , mask_token=lowerCamelCase , **lowerCamelCase , ) __a = vocab_file __a = False if not self.vocab_file else True def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __a = [self.cls_token_id] __a = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCamelCase ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __a = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase ): copyfile(self.vocab_file , lowerCamelCase ) return (out_vocab_file,)
268
0
'''simple docstring''' from __future__ import annotations def _A ( A__ , A__ ): """simple docstring""" if partitions <= 0: raise ValueError('''partitions must be a positive number!''' ) if partitions > number_of_bytes: raise ValueError('''partitions can not > number_of_bytes!''' ) __lowercase = number_of_bytes // partitions __lowercase = [] for i in range(SCREAMING_SNAKE_CASE_ ): __lowercase = i * bytes_per_partition + 1 __lowercase = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(F"{start_bytes}-{end_bytes}" ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
104
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionAttendAndExcitePipeline, UNetaDConditionModel, ) from diffusers.utils import load_numpy, skip_mps, slow from diffusers.utils.testing_utils import require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin lowerCamelCase__ = False @skip_mps class A__ ( __magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = StableDiffusionAttendAndExcitePipeline lowercase = False lowercase = TEXT_TO_IMAGE_PARAMS lowercase = TEXT_TO_IMAGE_BATCH_PARAMS.union({'token_indices'} ) lowercase = TEXT_TO_IMAGE_IMAGE_PARAMS lowercase = TEXT_TO_IMAGE_IMAGE_PARAMS @classmethod def _lowerCamelCase ( cls : Tuple ): '''simple docstring''' super().setUpClass() torch.use_deterministic_algorithms(a ) @classmethod def _lowerCamelCase ( cls : Any ): '''simple docstring''' super().tearDownClass() torch.use_deterministic_algorithms(a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase__ : Any = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=a , ) lowerCAmelCase__ : Any = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='scaled_linear' , clip_sample=a , set_alpha_to_one=a , ) torch.manual_seed(0 ) lowerCAmelCase__ : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) lowerCAmelCase__ : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='gelu' , projection_dim=512 , ) lowerCAmelCase__ : str = CLIPTextModel(a ) lowerCAmelCase__ : Optional[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowerCAmelCase__ : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _lowerCamelCase ( self : Union[str, Any] , a : Tuple , a : Union[str, Any]=0 ): '''simple docstring''' if str(a ).startswith('mps' ): lowerCAmelCase__ : List[str] = torch.manual_seed(a ) else: lowerCAmelCase__ : Any = torch.Generator(device=a ).manual_seed(a ) lowerCAmelCase__ : Optional[int] = { 'prompt': 'a cat and a frog', 'token_indices': [2, 5], 'generator': generator, 'num_inference_steps': 1, 'guidance_scale': 6.0, 'output_type': 'numpy', 'max_iter_to_alter': 2, 'thresholds': {0: 0.7}, } return inputs def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Tuple = 'cpu' lowerCAmelCase__ : Optional[int] = self.get_dummy_components() lowerCAmelCase__ : Dict = self.pipeline_class(**a ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) lowerCAmelCase__ : Union[str, Any] = self.get_dummy_inputs(a ) lowerCAmelCase__ : Union[str, Any] = pipe(**a ).images lowerCAmelCase__ : Optional[int] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 64, 64, 3) ) lowerCAmelCase__ : Dict = np.array( [0.6_3_9_0_5_3_6_4, 0.6_2_8_9_7_3_0_7, 0.4_8_5_9_9_0_1_7, 0.5_1_3_3_6_2_4, 0.5_5_5_0_0_4_8, 0.4_5_7_6_9_5_1_6, 0.5_0_3_2_6_9_7_3, 0.5_0_2_3_1_3_9, 0.4_5_3_8_4_4_9_6] ) lowerCAmelCase__ : Optional[int] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(a , 1E-3 ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' super().test_cpu_offload_forward_pass(expected_max_diff=5E-4 ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' self._test_inference_batch_single_identical(batch_size=2 , expected_max_diff=7E-4 ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' super().test_pt_np_pil_outputs_equivalent(expected_max_diff=5E-4 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' super().test_save_load_local(expected_max_difference=5E-4 ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=4E-4 ) @require_torch_gpu @slow class A__ ( unittest.TestCase ): @classmethod def _lowerCamelCase ( cls : List[str] ): '''simple docstring''' super().setUpClass() torch.use_deterministic_algorithms(a ) @classmethod def _lowerCamelCase ( cls : List[str] ): '''simple docstring''' super().tearDownClass() torch.use_deterministic_algorithms(a ) def _lowerCamelCase ( self : int ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : List[str] = torch.manual_seed(51 ) lowerCAmelCase__ : Any = StableDiffusionAttendAndExcitePipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , safety_checker=a , torch_dtype=torch.floataa ) pipe.to('cuda' ) lowerCAmelCase__ : Optional[int] = 'a painting of an elephant with glasses' lowerCAmelCase__ : Any = [5, 7] lowerCAmelCase__ : Optional[Any] = pipe( prompt=a , token_indices=a , guidance_scale=7.5 , generator=a , num_inference_steps=5 , max_iter_to_alter=5 , output_type='numpy' , ).images[0] lowerCAmelCase__ : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/attend-and-excite/elephant_glasses.npy' ) assert np.abs((expected_image - image).max() ) < 5E-1
212
0
from ....configuration_utils import PretrainedConfig from ....utils import logging lowerCamelCase : Tuple =logging.get_logger(__name__) lowerCamelCase : Optional[Any] ={ "speechbrain/m-ctc-t-large": "https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json", # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class __a ( UpperCAmelCase__ ): _lowerCAmelCase : Any = 'mctct' def __init__( self : str , SCREAMING_SNAKE_CASE : List[Any]=80_65 , SCREAMING_SNAKE_CASE : Any=15_36 , SCREAMING_SNAKE_CASE : Any=36 , SCREAMING_SNAKE_CASE : Optional[Any]=61_44 , SCREAMING_SNAKE_CASE : Optional[int]=4 , SCREAMING_SNAKE_CASE : Optional[int]=3_84 , SCREAMING_SNAKE_CASE : List[Any]=9_20 , SCREAMING_SNAKE_CASE : Dict=1e-5 , SCREAMING_SNAKE_CASE : Any=0.3 , SCREAMING_SNAKE_CASE : List[Any]="relu" , SCREAMING_SNAKE_CASE : str=0.0_2 , SCREAMING_SNAKE_CASE : List[str]=0.3 , SCREAMING_SNAKE_CASE : Union[str, Any]=0.3 , SCREAMING_SNAKE_CASE : Dict=1 , SCREAMING_SNAKE_CASE : Tuple=0 , SCREAMING_SNAKE_CASE : Union[str, Any]=2 , SCREAMING_SNAKE_CASE : Dict=1 , SCREAMING_SNAKE_CASE : Tuple=0.3 , SCREAMING_SNAKE_CASE : Any=1 , SCREAMING_SNAKE_CASE : Optional[Any]=(7,) , SCREAMING_SNAKE_CASE : Tuple=(3,) , SCREAMING_SNAKE_CASE : List[str]=80 , SCREAMING_SNAKE_CASE : Union[str, Any]=1 , SCREAMING_SNAKE_CASE : List[Any]=None , SCREAMING_SNAKE_CASE : Dict="sum" , SCREAMING_SNAKE_CASE : List[Any]=False , **SCREAMING_SNAKE_CASE : Union[str, Any] , ): '''simple docstring''' super().__init__(**_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE ) UpperCamelCase__ : int = vocab_size UpperCamelCase__ : str = hidden_size UpperCamelCase__ : str = num_hidden_layers UpperCamelCase__ : Dict = intermediate_size UpperCamelCase__ : List[str] = num_attention_heads UpperCamelCase__ : List[Any] = attention_head_dim UpperCamelCase__ : Union[str, Any] = max_position_embeddings UpperCamelCase__ : Optional[int] = layer_norm_eps UpperCamelCase__ : int = layerdrop UpperCamelCase__ : Tuple = hidden_act UpperCamelCase__ : Optional[int] = initializer_range UpperCamelCase__ : str = hidden_dropout_prob UpperCamelCase__ : Optional[int] = attention_probs_dropout_prob UpperCamelCase__ : Dict = pad_token_id UpperCamelCase__ : List[Any] = bos_token_id UpperCamelCase__ : List[str] = eos_token_id UpperCamelCase__ : Optional[int] = conv_glu_dim UpperCamelCase__ : Tuple = conv_dropout UpperCamelCase__ : Dict = num_conv_layers UpperCamelCase__ : Tuple = input_feat_per_channel UpperCamelCase__ : int = input_channels UpperCamelCase__ : str = conv_channels UpperCamelCase__ : Any = ctc_loss_reduction UpperCamelCase__ : Tuple = ctc_zero_infinity # prevents config testing fail with exporting to json UpperCamelCase__ : Optional[Any] = list(_SCREAMING_SNAKE_CASE ) UpperCamelCase__ : str = list(_SCREAMING_SNAKE_CASE ) if len(self.conv_kernel ) != self.num_conv_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.conv_kernel)` == `config.num_conv_layers` " F'but is `len(config.conv_kernel) = {len(self.conv_kernel )}`, ' F'`config.num_conv_layers = {self.num_conv_layers}`.' )
360
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCamelCase : List[Any] =logging.get_logger(__name__) lowerCamelCase : Optional[Any] ={ '''google/bit-50''': '''https://huggingface.co/google/bit-50/resolve/main/config.json''', } class __a ( A__ , A__ ): _lowerCAmelCase : Union[str, Any] = '''bit''' _lowerCAmelCase : List[str] = ['''preactivation''', '''bottleneck'''] _lowerCAmelCase : Any = ['''SAME''', '''VALID'''] def __init__( self : str , SCREAMING_SNAKE_CASE : Tuple=3 , SCREAMING_SNAKE_CASE : str=64 , SCREAMING_SNAKE_CASE : List[Any]=[2_56, 5_12, 10_24, 20_48] , SCREAMING_SNAKE_CASE : Union[str, Any]=[3, 4, 6, 3] , SCREAMING_SNAKE_CASE : str="preactivation" , SCREAMING_SNAKE_CASE : Any="relu" , SCREAMING_SNAKE_CASE : Dict=None , SCREAMING_SNAKE_CASE : List[Any]=32 , SCREAMING_SNAKE_CASE : Tuple=0.0 , SCREAMING_SNAKE_CASE : Dict=False , SCREAMING_SNAKE_CASE : int=32 , SCREAMING_SNAKE_CASE : str=1 , SCREAMING_SNAKE_CASE : List[str]=None , SCREAMING_SNAKE_CASE : Optional[int]=None , **SCREAMING_SNAKE_CASE : int , ): '''simple docstring''' super().__init__(**SCREAMING_SNAKE_CASE ) if layer_type not in self.layer_types: raise ValueError(F'layer_type={layer_type} is not one of {",".join(self.layer_types )}' ) if global_padding is not None: if global_padding.upper() in self.supported_padding: UpperCamelCase__ : Any = global_padding.upper() else: raise ValueError(F'Padding strategy {global_padding} not supported' ) UpperCamelCase__ : Dict = num_channels UpperCamelCase__ : Dict = embedding_size UpperCamelCase__ : Tuple = hidden_sizes UpperCamelCase__ : Any = depths UpperCamelCase__ : Optional[int] = layer_type UpperCamelCase__ : int = hidden_act UpperCamelCase__ : str = global_padding UpperCamelCase__ : Any = num_groups UpperCamelCase__ : str = drop_path_rate UpperCamelCase__ : Optional[Any] = embedding_dynamic_padding UpperCamelCase__ : Tuple = output_stride UpperCamelCase__ : List[str] = width_factor UpperCamelCase__ : Any = ["stem"] + [F'stage{idx}' for idx in range(1 , len(SCREAMING_SNAKE_CASE ) + 1 )] UpperCamelCase__ , UpperCamelCase__ : Union[str, Any] = get_aligned_output_features_output_indices( out_features=SCREAMING_SNAKE_CASE , out_indices=SCREAMING_SNAKE_CASE , stage_names=self.stage_names )
196
0
'''simple docstring''' import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = ['''image_processor''', '''tokenizer'''] snake_case = '''AutoImageProcessor''' snake_case = '''AutoTokenizer''' def __init__( self : Union[str, Any] , __UpperCAmelCase : Union[str, Any]=None , __UpperCAmelCase : str=None , **__UpperCAmelCase : Union[str, Any] ): '''simple docstring''' _A = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __UpperCAmelCase , ) _A = kwargs.pop("feature_extractor" ) _A = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__UpperCAmelCase , __UpperCAmelCase ) _A = self.image_processor _A = False def __call__( self : int , *__UpperCAmelCase : Dict , **__UpperCAmelCase : Dict ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*__UpperCAmelCase , **__UpperCAmelCase ) _A = kwargs.pop("images" , __UpperCAmelCase ) _A = kwargs.pop("text" , __UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: _A = args[0] _A = args[1:] if images is None and text is None: raise ValueError("You need to specify either an `images` or `text` input to process." ) if images is not None: _A = self.image_processor(__UpperCAmelCase , *__UpperCAmelCase , **__UpperCAmelCase ) if text is not None: _A = self.tokenizer(__UpperCAmelCase , **__UpperCAmelCase ) if text is None: return inputs elif images is None: return encodings else: _A = encodings["input_ids"] return inputs def lowerCAmelCase ( self : List[str] , *__UpperCAmelCase : Tuple , **__UpperCAmelCase : Any ): '''simple docstring''' return self.tokenizer.batch_decode(*__UpperCAmelCase , **__UpperCAmelCase ) def lowerCAmelCase ( self : Tuple , *__UpperCAmelCase : str , **__UpperCAmelCase : Dict ): '''simple docstring''' return self.tokenizer.decode(*__UpperCAmelCase , **__UpperCAmelCase ) @contextmanager def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' warnings.warn( "`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your " "labels by using the argument `text` of the regular `__call__` method (either in the same call as " "your images inputs, or in a separate call." ) _A = True _A = self.tokenizer yield _A = self.image_processor _A = False def lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Optional[int]=False , __UpperCAmelCase : str=None ): '''simple docstring''' if added_vocab is None: _A = self.tokenizer.get_added_vocab() _A = {} while tokens: _A = re.search(R"<s_(.*?)>" , __UpperCAmelCase , re.IGNORECASE ) if start_token is None: break _A = start_token.group(1 ) _A = re.search(Rf'''</s_{key}>''' , __UpperCAmelCase , re.IGNORECASE ) _A = start_token.group() if end_token is None: _A = tokens.replace(__UpperCAmelCase , "" ) else: _A = end_token.group() _A = re.escape(__UpperCAmelCase ) _A = re.escape(__UpperCAmelCase ) _A = re.search(f'''{start_token_escaped}(.*?){end_token_escaped}''' , __UpperCAmelCase , re.IGNORECASE ) if content is not None: _A = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _A = self.tokenajson(__UpperCAmelCase , is_inner_value=__UpperCAmelCase , added_vocab=__UpperCAmelCase ) if value: if len(__UpperCAmelCase ) == 1: _A = value[0] _A = value else: # leaf nodes _A = [] for leaf in content.split(R"<sep/>" ): _A = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _A = leaf[1:-2] # for categorical special tokens output[key].append(__UpperCAmelCase ) if len(output[key] ) == 1: _A = output[key][0] _A = tokens[tokens.find(__UpperCAmelCase ) + len(__UpperCAmelCase ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=__UpperCAmelCase , added_vocab=__UpperCAmelCase ) if len(__UpperCAmelCase ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def lowerCAmelCase ( self : Tuple ): '''simple docstring''' warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , __UpperCAmelCase , ) return self.image_processor_class @property def lowerCAmelCase ( self : List[str] ): '''simple docstring''' warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , __UpperCAmelCase , ) return self.image_processor
79
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase_ = { '''configuration_longformer''': [ '''LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongformerConfig''', '''LongformerOnnxConfig''', ], '''tokenization_longformer''': ['''LongformerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = ['''LongformerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongformerForMaskedLM''', '''LongformerForMultipleChoice''', '''LongformerForQuestionAnswering''', '''LongformerForSequenceClassification''', '''LongformerForTokenClassification''', '''LongformerModel''', '''LongformerPreTrainedModel''', '''LongformerSelfAttention''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLongformerForMaskedLM''', '''TFLongformerForMultipleChoice''', '''TFLongformerForQuestionAnswering''', '''TFLongformerForSequenceClassification''', '''TFLongformerForTokenClassification''', '''TFLongformerModel''', '''TFLongformerPreTrainedModel''', '''TFLongformerSelfAttention''', ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
79
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCamelCase = { '''configuration_bloom''': ['''BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BloomConfig''', '''BloomOnnxConfig'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = ['''BloomTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ '''BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BloomForCausalLM''', '''BloomModel''', '''BloomPreTrainedModel''', '''BloomForSequenceClassification''', '''BloomForTokenClassification''', '''BloomForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys __lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
101
'''simple docstring''' from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, ) -> list[float]: A_ , A_ = coefficient_matrix.shape A_ , A_ = constant_matrix.shape if rowsa != colsa: A_ = F'''Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}''' raise ValueError(UpperCAmelCase__ ) if colsa != 1: A_ = F'''Constant matrix must be nx1 but received {rowsa}x{colsa}''' raise ValueError(UpperCAmelCase__ ) if rowsa != rowsa: A_ = ( """Coefficient and constant matrices dimensions must be nxn and nx1 but """ F'''received {rowsa}x{colsa} and {rowsa}x{colsa}''' ) raise ValueError(UpperCAmelCase__ ) if len(UpperCAmelCase__ ) != rowsa: A_ = ( """Number of initial values must be equal to number of rows in coefficient """ F'''matrix but received {len(UpperCAmelCase__ )} and {rowsa}''' ) raise ValueError(UpperCAmelCase__ ) if iterations <= 0: raise ValueError("""Iterations must be at least 1""" ) A_ = np.concatenate( (coefficient_matrix, constant_matrix), axis=1 ) A_ , A_ = table.shape strictly_diagonally_dominant(UpperCAmelCase__ ) # Iterates the whole matrix for given number of times for _ in range(UpperCAmelCase__ ): A_ = [] for row in range(UpperCAmelCase__ ): A_ = 0 for col in range(UpperCAmelCase__ ): if col == row: A_ = table[row][col] elif col == cols - 1: A_ = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] A_ = (temp + val) / denom new_val.append(UpperCAmelCase__ ) A_ = new_val return [float(UpperCAmelCase__ ) for i in new_val] def UpperCAmelCase__ ( UpperCAmelCase__ ) -> bool: A_ , A_ = table.shape A_ = True for i in range(0, UpperCAmelCase__ ): A_ = 0 for j in range(0, cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("""Coefficient matrix is not strictly diagonally dominant""" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
101
1