code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''microsoft/markuplm-base''': '''https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json''', '''microsoft/markuplm-large''': '''https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json''', } class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = '''markuplm''' def __init__( self : Dict , __UpperCAmelCase : Union[str, Any]=30522 , __UpperCAmelCase : Tuple=768 , __UpperCAmelCase : Any=12 , __UpperCAmelCase : List[str]=12 , __UpperCAmelCase : str=3072 , __UpperCAmelCase : List[Any]="gelu" , __UpperCAmelCase : List[str]=0.1 , __UpperCAmelCase : List[Any]=0.1 , __UpperCAmelCase : Union[str, Any]=512 , __UpperCAmelCase : List[str]=2 , __UpperCAmelCase : int=0.02 , __UpperCAmelCase : List[str]=1E-12 , __UpperCAmelCase : Optional[int]=0 , __UpperCAmelCase : str=0 , __UpperCAmelCase : Tuple=2 , __UpperCAmelCase : Tuple=256 , __UpperCAmelCase : int=1024 , __UpperCAmelCase : Union[str, Any]=216 , __UpperCAmelCase : List[str]=1001 , __UpperCAmelCase : Any=32 , __UpperCAmelCase : Tuple=50 , __UpperCAmelCase : str="absolute" , __UpperCAmelCase : Tuple=True , __UpperCAmelCase : List[Any]=None , **__UpperCAmelCase : Union[str, Any] , ): '''simple docstring''' super().__init__( pad_token_id=__snake_case , bos_token_id=__snake_case , eos_token_id=__snake_case , **__snake_case , ) _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = hidden_act _A = intermediate_size _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = initializer_range _A = layer_norm_eps _A = position_embedding_type _A = use_cache _A = classifier_dropout # additional properties _A = max_depth _A = max_xpath_tag_unit_embeddings _A = max_xpath_subs_unit_embeddings _A = tag_pad_id _A = subs_pad_id _A = xpath_unit_hidden_size
79
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase ( unittest.TestCase ): @slow def lowercase__ ( self : List[Any] ) -> str: _lowerCAmelCase = AutoModelForSeqaSeqLM.from_pretrained("""google/mt5-small""" , return_dict=__snake_case ).to(__snake_case ) _lowerCAmelCase = AutoTokenizer.from_pretrained("""google/mt5-small""" ) _lowerCAmelCase = tokenizer("""Hello there""" , return_tensors="""pt""" ).input_ids _lowerCAmelCase = tokenizer("""Hi I am""" , return_tensors="""pt""" ).input_ids _lowerCAmelCase = model(input_ids.to(__snake_case ) , labels=labels.to(__snake_case ) ).loss _lowerCAmelCase = -(labels.shape[-1] * loss.item()) _lowerCAmelCase = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
70
0
import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig A_ : List[str] = { '''facebook/maskformer-swin-base-ade''': ( '''https://huggingface.co/facebook/maskformer-swin-base-ade/blob/main/config.json''' ) # See all MaskFormer models at https://huggingface.co/models?filter=maskformer } A_ : List[str] = logging.get_logger(__name__) class A_ ( snake_case_ ): '''simple docstring''' a__ = '''maskformer''' a__ = {'''hidden_size''': '''mask_feature_size'''} a__ = ['''resnet''', '''swin'''] a__ = ['''detr'''] def __init__(self , lowercase__ = 256 , lowercase__ = 256 , lowercase__ = 0.1 , lowercase__ = False , lowercase__ = None , lowercase__ = None , lowercase__ = 0.02 , lowercase__ = 1.0 , lowercase__ = 1.0 , lowercase__ = 1.0 , lowercase__ = 20.0 , lowercase__ = None , **lowercase__ , ) -> int: if backbone_config is None: # fall back to https://huggingface.co/microsoft/swin-base-patch4-window12-384-in22k __UpperCAmelCase = SwinConfig( image_size=384 , in_channels=3 , patch_size=4 , embed_dim=128 , depths=[2, 2, 18, 2] , num_heads=[4, 8, 16, 32] , window_size=12 , drop_path_rate=0.3 , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] , ) if isinstance(__snake_case , __snake_case ): __UpperCAmelCase = backbone_config.pop('''model_type''' ) __UpperCAmelCase = CONFIG_MAPPING[backbone_model_type] __UpperCAmelCase = config_class.from_dict(__snake_case ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F'''Backbone {backbone_config.model_type} is not a supported model and may not be compatible with MaskFormer. ''' F'''Supported model types: {','.join(self.backbones_supported )}''' ) if decoder_config is None: # fall back to https://huggingface.co/facebook/detr-resnet-50 __UpperCAmelCase = DetrConfig() else: # verify that the decoder is supported __UpperCAmelCase = ( decoder_config.pop('''model_type''' ) if isinstance(__snake_case , __snake_case ) else decoder_config.model_type ) if decoder_type not in self.decoders_supported: raise ValueError( F'''Transformer Decoder {decoder_type} not supported, please use one of''' F''' {','.join(self.decoders_supported )}''' ) if isinstance(__snake_case , __snake_case ): __UpperCAmelCase = CONFIG_MAPPING[decoder_type] __UpperCAmelCase = config_class.from_dict(__snake_case ) __UpperCAmelCase = backbone_config __UpperCAmelCase = decoder_config # main feature dimension for the model __UpperCAmelCase = fpn_feature_size __UpperCAmelCase = mask_feature_size # initializer __UpperCAmelCase = init_std __UpperCAmelCase = init_xavier_std # Hungarian matcher && loss __UpperCAmelCase = cross_entropy_weight __UpperCAmelCase = dice_weight __UpperCAmelCase = mask_weight __UpperCAmelCase = use_auxiliary_loss __UpperCAmelCase = no_object_weight __UpperCAmelCase = output_auxiliary_logits __UpperCAmelCase = self.decoder_config.encoder_attention_heads __UpperCAmelCase = self.decoder_config.num_hidden_layers super().__init__(**__snake_case ) @classmethod def lowerCAmelCase_ (cls , lowercase__ , lowercase__ , **lowercase__ ) -> Union[str, Any]: return cls( backbone_config=__snake_case , decoder_config=__snake_case , **__snake_case , ) def lowerCAmelCase_ (self ) -> Dict[str, any]: __UpperCAmelCase = copy.deepcopy(self.__dict__ ) __UpperCAmelCase = self.backbone_config.to_dict() __UpperCAmelCase = self.decoder_config.to_dict() __UpperCAmelCase = self.__class__.model_type return output
333
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging A__ : Any =logging.get_logger(__name__) A__ : List[Any] ='''▁''' A__ : Optional[int] ={'''vocab_file''': '''sentencepiece.bpe.model'''} A__ : Union[str, Any] ={ '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model''' ), } } A__ : Dict ={ '''facebook/nllb-200-distilled-600M''': 10_24, } # fmt: off A__ : Union[str, Any] =['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCAmelCase ( snake_case_ ): _lowercase: int = VOCAB_FILES_NAMES _lowercase: str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase: Dict = PRETRAINED_VOCAB_FILES_MAP _lowercase: str = ['''input_ids''', '''attention_mask'''] _lowercase: List[int] = [] _lowercase: List[int] = [] def __init__( self : int , __snake_case : Optional[Any] , __snake_case : Dict="<s>" , __snake_case : Optional[int]="</s>" , __snake_case : Dict="</s>" , __snake_case : str="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : List[Any]=None , __snake_case : Union[str, Any]=None , __snake_case : int=None , __snake_case : Optional[Dict[str, Any]] = None , __snake_case : str=None , __snake_case : str=False , **__snake_case : List[Any] , ) -> List[str]: # Mask token behave like a normal word, i.e. include the space before it _lowerCAmelCase = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token _lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase = legacy_behaviour super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , tokenizer_file=__snake_case , src_lang=__snake_case , tgt_lang=__snake_case , additional_special_tokens=__snake_case , sp_model_kwargs=self.sp_model_kwargs , legacy_behaviour=__snake_case , **__snake_case , ) _lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) _lowerCAmelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token _lowerCAmelCase = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _lowerCAmelCase = 1 _lowerCAmelCase = len(self.sp_model ) _lowerCAmelCase = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(__snake_case ) } _lowerCAmelCase = {v: k for k, v in self.lang_code_to_id.items()} _lowerCAmelCase = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) _lowerCAmelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} _lowerCAmelCase = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) _lowerCAmelCase = src_lang if src_lang is not None else """eng_Latn""" _lowerCAmelCase = self.lang_code_to_id[self._src_lang] _lowerCAmelCase = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : List[str] ) -> List[str]: _lowerCAmelCase = self.__dict__.copy() _lowerCAmelCase = None _lowerCAmelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self : Dict , __snake_case : Optional[Any] ) -> Dict: _lowerCAmelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _lowerCAmelCase = {} _lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def lowercase__ ( self : List[Any] ) -> Any: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def lowercase__ ( self : int ) -> str: return self._src_lang @src_lang.setter def lowercase__ ( self : Dict , __snake_case : str ) -> None: _lowerCAmelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase__ ( self : List[Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) _lowerCAmelCase = [1] * len(self.prefix_tokens ) _lowerCAmelCase = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__snake_case )) + suffix_ones return prefix_ones + ([0] * len(__snake_case )) + ([0] * len(__snake_case )) + suffix_ones def lowercase__ ( self : Optional[Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase__ ( self : Optional[Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: _lowerCAmelCase = [self.sep_token_id] _lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase__ ( self : List[str] , __snake_case : Union[str, Any] , __snake_case : str , __snake_case : Optional[str] , __snake_case : Optional[str] , **__snake_case : Optional[int] ) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _lowerCAmelCase = src_lang _lowerCAmelCase = self(__snake_case , add_special_tokens=__snake_case , return_tensors=__snake_case , **__snake_case ) _lowerCAmelCase = self.convert_tokens_to_ids(__snake_case ) _lowerCAmelCase = tgt_lang_id return inputs def lowercase__ ( self : List[Any] ) -> Optional[int]: _lowerCAmelCase = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase__ ( self : Optional[int] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def lowercase__ ( self : Optional[Any] , __snake_case : Union[str, Any] ) -> Tuple: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _lowerCAmelCase = self.sp_model.PieceToId(__snake_case ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowercase__ ( self : List[Any] , __snake_case : Union[str, Any] ) -> Optional[int]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowercase__ ( self : Optional[int] , __snake_case : Union[str, Any] ) -> str: _lowerCAmelCase = """""".join(__snake_case ).replace(__snake_case , """ """ ).strip() return out_string def lowercase__ ( self : str , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return _lowerCAmelCase = os.path.join( __snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , """wb""" ) as fi: _lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,) def lowercase__ ( self : Optional[Any] , __snake_case : List[str] , __snake_case : str = "eng_Latn" , __snake_case : Optional[List[str]] = None , __snake_case : str = "fra_Latn" , **__snake_case : Optional[int] , ) -> BatchEncoding: _lowerCAmelCase = src_lang _lowerCAmelCase = tgt_lang return super().prepare_seqaseq_batch(__snake_case , __snake_case , **__snake_case ) def lowercase__ ( self : str ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase__ ( self : Dict ) -> Optional[Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase__ ( self : str , __snake_case : int ) -> None: _lowerCAmelCase = self.lang_code_to_id[src_lang] if self.legacy_behaviour: _lowerCAmelCase = [] _lowerCAmelCase = [self.eos_token_id, self.cur_lang_code] else: _lowerCAmelCase = [self.cur_lang_code] _lowerCAmelCase = [self.eos_token_id] def lowercase__ ( self : Any , __snake_case : str ) -> None: _lowerCAmelCase = self.lang_code_to_id[lang] if self.legacy_behaviour: _lowerCAmelCase = [] _lowerCAmelCase = [self.eos_token_id, self.cur_lang_code] else: _lowerCAmelCase = [self.cur_lang_code] _lowerCAmelCase = [self.eos_token_id]
70
0
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor __A : str = logging.get_logger(__name__) class A_ (snake_case_ ): def __init__( self , *_A , **_A ): '''simple docstring''' warnings.warn( '''The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use GLPNImageProcessor instead.''' , __snake_case , ) super().__init__(*__snake_case , **__snake_case )
273
'''simple docstring''' def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = len(lowerCAmelCase ) for i in range(length - 1 ): _lowerCAmelCase = i for k in range(i + 1 , lowerCAmelCase ): if collection[k] < collection[least]: _lowerCAmelCase = k if least != i: _lowerCAmelCase , _lowerCAmelCase = (collection[i], collection[least]) return collection if __name__ == "__main__": A__ : str =input('''Enter numbers separated by a comma:\n''').strip() A__ : Optional[int] =[int(item) for item in user_input.split(''',''')] print(selection_sort(unsorted))
70
0
'''simple docstring''' from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def a_ ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase=1E-12 ) -> Union[str, Any]: __lowerCamelCase : Any = jnp.divide(emb_a.T ,jnp.clip(jnp.linalg.norm(_lowerCAmelCase ,axis=1 ) ,a_min=_lowerCAmelCase ) ).T __lowerCamelCase : str = jnp.divide(emb_a.T ,jnp.clip(jnp.linalg.norm(_lowerCAmelCase ,axis=1 ) ,a_min=_lowerCAmelCase ) ).T return jnp.matmul(_lowerCAmelCase ,norm_emb_a.T ) class lowerCamelCase_ ( nn.Module ): """simple docstring""" a_ =42 a_ =jnp.floataa def _lowercase ( self : Optional[int] ) -> List[str]: __lowerCamelCase : Tuple = FlaxCLIPVisionModule(self.config.vision_config ) __lowerCamelCase : Optional[Any] = nn.Dense(self.config.projection_dim , use_bias=__snake_case , dtype=self.dtype ) __lowerCamelCase : Any = self.param('concept_embeds' , jax.nn.initializers.ones , (17, self.config.projection_dim) ) __lowerCamelCase : List[str] = self.param( 'special_care_embeds' , jax.nn.initializers.ones , (3, self.config.projection_dim) ) __lowerCamelCase : Any = self.param('concept_embeds_weights' , jax.nn.initializers.ones , (17,) ) __lowerCamelCase : Tuple = self.param('special_care_embeds_weights' , jax.nn.initializers.ones , (3,) ) def __call__( self : Optional[Any] , _a : Tuple ) -> int: __lowerCamelCase : List[Any] = self.vision_model(__snake_case )[1] __lowerCamelCase : List[Any] = self.visual_projection(__snake_case ) __lowerCamelCase : List[str] = jax_cosine_distance(__snake_case , self.special_care_embeds ) __lowerCamelCase : Optional[Any] = jax_cosine_distance(__snake_case , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs __lowerCamelCase : Any = 0.0 __lowerCamelCase : int = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment __lowerCamelCase : int = jnp.round(__snake_case , 3 ) __lowerCamelCase : Optional[int] = jnp.any(special_scores > 0 , axis=1 , keepdims=__snake_case ) # Use a lower threshold if an image has any special care concept __lowerCamelCase : Optional[int] = is_special_care * 0.01 __lowerCamelCase : List[str] = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment __lowerCamelCase : str = jnp.round(__snake_case , 3 ) __lowerCamelCase : Union[str, Any] = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class lowerCamelCase_ ( snake_case_ ): """simple docstring""" a_ =CLIPConfig a_ ='''clip_input''' a_ =FlaxStableDiffusionSafetyCheckerModule def __init__( self : List[str] , _a : CLIPConfig , _a : Optional[Tuple] = None , _a : int = 0 , _a : jnp.dtype = jnp.floataa , _a : bool = True , **_a : str , ) -> List[str]: if input_shape is None: __lowerCamelCase : Dict = (1, 224, 224, 3) __lowerCamelCase : Optional[Any] = self.module_class(config=__snake_case , dtype=__snake_case , **__snake_case ) super().__init__(__snake_case , __snake_case , input_shape=__snake_case , seed=__snake_case , dtype=__snake_case , _do_init=_do_init ) def _lowercase ( self : str , _a : jax.random.KeyArray , _a : Tuple , _a : FrozenDict = None ) -> FrozenDict: # init input tensor __lowerCamelCase : str = jax.random.normal(__snake_case , __snake_case ) __lowerCamelCase ,__lowerCamelCase : str = jax.random.split(__snake_case ) __lowerCamelCase : Union[str, Any] = {'params': params_rng, 'dropout': dropout_rng} __lowerCamelCase : Dict = self.module.init(__snake_case , __snake_case )['params'] return random_params def __call__( self : Any , _a : Optional[int] , _a : dict = None , ) -> int: __lowerCamelCase : Dict = jnp.transpose(__snake_case , (0, 2, 3, 1) ) return self.module.apply( {'params': params or self.params} , jnp.array(__snake_case , dtype=jnp.floataa ) , rngs={} , )
208
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL A__ : List[str] =logging.get_logger(__name__) def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if isinstance(lowerCAmelCase , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowerCAmelCase , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowerCAmelCase ): return [[videos]] raise ValueError(f"Could not make batched video from {videos}" ) class UpperCAmelCase ( snake_case_ ): _lowercase: Any = ['''pixel_values'''] def __init__( self : Tuple , __snake_case : bool = True , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = PILImageResampling.BILINEAR , __snake_case : bool = True , __snake_case : Dict[str, int] = None , __snake_case : bool = True , __snake_case : Union[int, float] = 1 / 2_55 , __snake_case : bool = True , __snake_case : bool = True , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , **__snake_case : str , ) -> None: super().__init__(**__snake_case ) _lowerCAmelCase = size if size is not None else {"""shortest_edge""": 2_56} _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) _lowerCAmelCase = crop_size if crop_size is not None else {"""height""": 2_24, """width""": 2_24} _lowerCAmelCase = get_size_dict(__snake_case , param_name="""crop_size""" ) _lowerCAmelCase = do_resize _lowerCAmelCase = size _lowerCAmelCase = do_center_crop _lowerCAmelCase = crop_size _lowerCAmelCase = resample _lowerCAmelCase = do_rescale _lowerCAmelCase = rescale_factor _lowerCAmelCase = offset _lowerCAmelCase = do_normalize _lowerCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase__ ( self : int , __snake_case : np.ndarray , __snake_case : Dict[str, int] , __snake_case : PILImageResampling = PILImageResampling.BILINEAR , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Optional[Any] , ) -> np.ndarray: _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) if "shortest_edge" in size: _lowerCAmelCase = get_resize_output_image_size(__snake_case , size["""shortest_edge"""] , default_to_square=__snake_case ) elif "height" in size and "width" in size: _lowerCAmelCase = (size["""height"""], size["""width"""]) else: raise ValueError(f"Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}" ) return resize(__snake_case , size=__snake_case , resample=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Union[str, Any] , __snake_case : np.ndarray , __snake_case : Dict[str, int] , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : List[Any] , ) -> np.ndarray: _lowerCAmelCase = get_size_dict(__snake_case ) if "height" not in size or "width" not in size: raise ValueError(f"Size must have 'height' and 'width' as keys. Got {size.keys()}" ) return center_crop(__snake_case , size=(size["""height"""], size["""width"""]) , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Union[str, Any] , __snake_case : np.ndarray , __snake_case : Union[int, float] , __snake_case : bool = True , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Optional[Any] , ) -> Dict: _lowerCAmelCase = image.astype(np.floataa ) if offset: _lowerCAmelCase = image - (scale / 2) return rescale(__snake_case , scale=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Optional[int] , __snake_case : np.ndarray , __snake_case : Union[float, List[float]] , __snake_case : Union[float, List[float]] , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Tuple , ) -> np.ndarray: return normalize(__snake_case , mean=__snake_case , std=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : List[Any] , __snake_case : ImageInput , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = None , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : bool = None , __snake_case : float = None , __snake_case : bool = None , __snake_case : bool = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) if offset and not do_rescale: raise ValueError("""For offset, do_rescale must also be set to True.""" ) # All transformations expect numpy arrays. _lowerCAmelCase = to_numpy_array(__snake_case ) if do_resize: _lowerCAmelCase = self.resize(image=__snake_case , size=__snake_case , resample=__snake_case ) if do_center_crop: _lowerCAmelCase = self.center_crop(__snake_case , size=__snake_case ) if do_rescale: _lowerCAmelCase = self.rescale(image=__snake_case , scale=__snake_case , offset=__snake_case ) if do_normalize: _lowerCAmelCase = self.normalize(image=__snake_case , mean=__snake_case , std=__snake_case ) _lowerCAmelCase = to_channel_dimension_format(__snake_case , __snake_case ) return image def lowercase__ ( self : List[Any] , __snake_case : ImageInput , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = None , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : bool = None , __snake_case : float = None , __snake_case : bool = None , __snake_case : bool = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[str, TensorType]] = None , __snake_case : ChannelDimension = ChannelDimension.FIRST , **__snake_case : List[str] , ) -> PIL.Image.Image: _lowerCAmelCase = do_resize if do_resize is not None else self.do_resize _lowerCAmelCase = resample if resample is not None else self.resample _lowerCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop _lowerCAmelCase = do_rescale if do_rescale is not None else self.do_rescale _lowerCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCAmelCase = offset if offset is not None else self.offset _lowerCAmelCase = do_normalize if do_normalize is not None else self.do_normalize _lowerCAmelCase = image_mean if image_mean is not None else self.image_mean _lowerCAmelCase = image_std if image_std is not None else self.image_std _lowerCAmelCase = size if size is not None else self.size _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) _lowerCAmelCase = crop_size if crop_size is not None else self.crop_size _lowerCAmelCase = get_size_dict(__snake_case , param_name="""crop_size""" ) if not valid_images(__snake_case ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) _lowerCAmelCase = make_batched(__snake_case ) _lowerCAmelCase = [ [ self._preprocess_image( image=__snake_case , do_resize=__snake_case , size=__snake_case , resample=__snake_case , do_center_crop=__snake_case , crop_size=__snake_case , do_rescale=__snake_case , rescale_factor=__snake_case , offset=__snake_case , do_normalize=__snake_case , image_mean=__snake_case , image_std=__snake_case , data_format=__snake_case , ) for img in video ] for video in videos ] _lowerCAmelCase = {"""pixel_values""": videos} return BatchFeature(data=__snake_case , tensor_type=__snake_case )
70
0
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def a ( __a , __a , __a=None ) -> Union[str, Any]: '''simple docstring''' assert torch_layer.weight.shape == weight.shape, f'''{torch_layer} layer.weight does not match''' UpperCamelCase__ :int = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f'''{torch_layer} layer.bias does not match''' UpperCamelCase__ :int = nn.Parameter(__a ) def a ( __a , __a , __a ) -> str: '''simple docstring''' UpperCamelCase__ :Dict = np.asarray(weights[0] ) UpperCamelCase__ :Tuple = np.asarray(weights[1] ) UpperCamelCase__ :List[Any] = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def a ( __a , __a , __a ) -> List[str]: '''simple docstring''' UpperCamelCase__ :str = np.asarray(weights[0] ) UpperCamelCase__ :Optional[Any] = np.asarray(weights[1] ) UpperCamelCase__ :Any = np.asarray(weights[2] ) UpperCamelCase__ :Optional[int] = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def a ( __a , __a , __a ) -> Tuple: '''simple docstring''' UpperCamelCase__ :List[str] = weights[0][0][0] UpperCamelCase__ :str = np.asarray(layer_norm_a[0] ) UpperCamelCase__ :Optional[Any] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output UpperCamelCase__ :int = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs UpperCamelCase__ :str = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: UpperCamelCase__ :int = intermediate_weights[2] # layernorm 2 UpperCamelCase__ :Optional[Any] = np.asarray(intermediate_weights[0][0] ) UpperCamelCase__ :Any = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense UpperCamelCase__ :Optional[Any] = np.asarray(intermediate_weights[1][0] ) UpperCamelCase__ :int = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out UpperCamelCase__ :Optional[int] = np.asarray(intermediate_weights[4][0] ) UpperCamelCase__ :Optional[Any] = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def a ( __a , __a , __a ) -> str: '''simple docstring''' UpperCamelCase__ :Dict = torch_model.reformer # word embeds UpperCamelCase__ :Optional[int] = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): UpperCamelCase__ :List[str] = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): UpperCamelCase__ :List[str] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f'''{position_embeddings[emb_idx]} emb does not match''' UpperCamelCase__ :List[Any] = nn.Parameter(torch.tensor(__a ) ) UpperCamelCase__ :Optional[int] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): UpperCamelCase__ :List[Any] = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm UpperCamelCase__ :Tuple = np.asarray(weights[7][0] ) UpperCamelCase__ :Optional[Any] = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings UpperCamelCase__ :Union[str, Any] = np.asarray(weights[9][0] ) UpperCamelCase__ :Union[str, Any] = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def a ( __a , __a , __a ) -> List[Any]: '''simple docstring''' UpperCamelCase__ :Dict = ReformerConfig.from_json_file(__a ) print(f'''Building PyTorch model from configuration: {config}''' ) UpperCamelCase__ :Optional[int] = ReformerModelWithLMHead(__a ) with open(__a , '''rb''' ) as f: UpperCamelCase__ :List[Any] = pickle.load(__a )['''weights'''] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __snake_case = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
97
'''simple docstring''' import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class UpperCAmelCase ( snake_case_ ): _lowercase: Union[str, Any] = ['''image_processor''', '''tokenizer'''] _lowercase: int = '''AutoImageProcessor''' _lowercase: Optional[int] = '''AutoTokenizer''' def __init__( self : int , __snake_case : Tuple=None , __snake_case : Optional[int]=None , **__snake_case : Tuple ) -> List[Any]: _lowerCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __snake_case , ) _lowerCAmelCase = kwargs.pop("""feature_extractor""" ) _lowerCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__snake_case , __snake_case ) _lowerCAmelCase = self.image_processor _lowerCAmelCase = False def __call__( self : Dict , *__snake_case : Optional[int] , **__snake_case : Union[str, Any] ) -> Tuple: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*__snake_case , **__snake_case ) _lowerCAmelCase = kwargs.pop("""images""" , __snake_case ) _lowerCAmelCase = kwargs.pop("""text""" , __snake_case ) if len(__snake_case ) > 0: _lowerCAmelCase = args[0] _lowerCAmelCase = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _lowerCAmelCase = self.image_processor(__snake_case , *__snake_case , **__snake_case ) if text is not None: _lowerCAmelCase = self.tokenizer(__snake_case , **__snake_case ) if text is None: return inputs elif images is None: return encodings else: _lowerCAmelCase = encodings["""input_ids"""] return inputs def lowercase__ ( self : List[Any] , *__snake_case : Dict , **__snake_case : List[str] ) -> int: return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def lowercase__ ( self : int , *__snake_case : Tuple , **__snake_case : Optional[Any] ) -> Any: return self.tokenizer.decode(*__snake_case , **__snake_case ) @contextmanager def lowercase__ ( self : int ) -> Optional[Any]: warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _lowerCAmelCase = True _lowerCAmelCase = self.tokenizer yield _lowerCAmelCase = self.image_processor _lowerCAmelCase = False def lowercase__ ( self : Optional[int] , __snake_case : Union[str, Any] , __snake_case : List[Any]=False , __snake_case : Dict=None ) -> Tuple: if added_vocab is None: _lowerCAmelCase = self.tokenizer.get_added_vocab() _lowerCAmelCase = {} while tokens: _lowerCAmelCase = re.search(R"""<s_(.*?)>""" , __snake_case , re.IGNORECASE ) if start_token is None: break _lowerCAmelCase = start_token.group(1 ) _lowerCAmelCase = re.search(Rf"</s_{key}>" , __snake_case , re.IGNORECASE ) _lowerCAmelCase = start_token.group() if end_token is None: _lowerCAmelCase = tokens.replace(__snake_case , """""" ) else: _lowerCAmelCase = end_token.group() _lowerCAmelCase = re.escape(__snake_case ) _lowerCAmelCase = re.escape(__snake_case ) _lowerCAmelCase = re.search(f"{start_token_escaped}(.*?){end_token_escaped}" , __snake_case , re.IGNORECASE ) if content is not None: _lowerCAmelCase = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _lowerCAmelCase = self.tokenajson(__snake_case , is_inner_value=__snake_case , added_vocab=__snake_case ) if value: if len(__snake_case ) == 1: _lowerCAmelCase = value[0] _lowerCAmelCase = value else: # leaf nodes _lowerCAmelCase = [] for leaf in content.split(R"""<sep/>""" ): _lowerCAmelCase = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _lowerCAmelCase = leaf[1:-2] # for categorical special tokens output[key].append(__snake_case ) if len(output[key] ) == 1: _lowerCAmelCase = output[key][0] _lowerCAmelCase = tokens[tokens.find(__snake_case ) + len(__snake_case ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=__snake_case , added_vocab=__snake_case ) if len(__snake_case ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def lowercase__ ( self : Optional[int] ) -> Union[str, Any]: warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , __snake_case , ) return self.image_processor_class @property def lowercase__ ( self : List[Any] ) -> Any: warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , __snake_case , ) return self.image_processor
70
0
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class __A ( snake_case_ ): """simple docstring""" __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = None class __A ( snake_case_ ): """simple docstring""" def __init__( self , __A=1 , __A=0 , __A=2 , __A=512 , __A="cls" , __A=False , __A=True , **__A , ) -> Union[str, Any]: super().__init__(pad_token_id=__snake_case , bos_token_id=__snake_case , eos_token_id=__snake_case , **__snake_case ) a =project_dim a =pooler_fn a =learn_encoder a =use_attention_mask class __A ( snake_case_ ): """simple docstring""" __lowerCAmelCase = [r'''pooler''', r'''logit_scale'''] __lowerCAmelCase = [r'''position_ids''', r'''predictions.decoder.bias'''] __lowerCAmelCase = '''roberta''' __lowerCAmelCase = RobertaSeriesConfig def __init__( self , __A ) -> Optional[Any]: super().__init__(__snake_case ) a =XLMRobertaModel(__snake_case ) a =nn.Linear(config.hidden_size , config.project_dim ) a =getattr(__snake_case , '''has_pre_transformation''' , __snake_case ) if self.has_pre_transformation: a =nn.Linear(config.hidden_size , config.project_dim ) a =nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def SCREAMING_SNAKE_CASE ( self , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , ) -> int: a =return_dict if return_dict is not None else self.config.use_return_dict a =self.base_model( input_ids=__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , position_ids=__snake_case , head_mask=__snake_case , inputs_embeds=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , output_attentions=__snake_case , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__snake_case , ) if self.has_pre_transformation: a =outputs['''hidden_states'''][-2] a =self.pre_LN(__snake_case ) a =self.transformation_pre(__snake_case ) return TransformationModelOutput( projection_state=__snake_case , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: a =self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__snake_case , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
'''simple docstring''' from __future__ import annotations import math def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True A__ : Optional[Any] =[num for num in range(3, 10_00_01, 2) if not is_prime(num)] def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if not isinstance(lowerCAmelCase , lowerCAmelCase ): raise ValueError("""n must be an integer""" ) if n <= 0: raise ValueError("""n must be >= 0""" ) _lowerCAmelCase = [] for num in range(len(lowerCAmelCase ) ): _lowerCAmelCase = 0 while 2 * i * i <= odd_composites[num]: _lowerCAmelCase = odd_composites[num] - 2 * i * i if is_prime(lowerCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(lowerCAmelCase ) == n: return list_nums return [] def UpperCamelCase__ ( ): """simple docstring""" return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
70
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available snake_case_ = { '''configuration_canine''': ['''CANINE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CanineConfig'''], '''tokenization_canine''': ['''CanineTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ '''CANINE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CanineForMultipleChoice''', '''CanineForQuestionAnswering''', '''CanineForSequenceClassification''', '''CanineForTokenClassification''', '''CanineLayer''', '''CanineModel''', '''CaninePreTrainedModel''', '''load_tf_weights_in_canine''', ] if TYPE_CHECKING: from .configuration_canine import CANINE_PRETRAINED_CONFIG_ARCHIVE_MAP, CanineConfig from .tokenization_canine import CanineTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_canine import ( CANINE_PRETRAINED_MODEL_ARCHIVE_LIST, CanineForMultipleChoice, CanineForQuestionAnswering, CanineForSequenceClassification, CanineForTokenClassification, CanineLayer, CanineModel, CaninePreTrainedModel, load_tf_weights_in_canine, ) else: import sys snake_case_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
78
'''simple docstring''' import argparse import json from tqdm import tqdm def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--src_path""" , type=lowerCAmelCase , default="""biencoder-nq-dev.json""" , help="""Path to raw DPR training data""" , ) parser.add_argument( """--evaluation_set""" , type=lowerCAmelCase , help="""where to store parsed evaluation_set file""" , ) parser.add_argument( """--gold_data_path""" , type=lowerCAmelCase , help="""where to store parsed gold_data_path file""" , ) _lowerCAmelCase = parser.parse_args() with open(args.src_path , """r""" ) as src_file, open(args.evaluation_set , """w""" ) as eval_file, open( args.gold_data_path , """w""" ) as gold_file: _lowerCAmelCase = json.load(lowerCAmelCase ) for dpr_record in tqdm(lowerCAmelCase ): _lowerCAmelCase = dpr_record["""question"""] _lowerCAmelCase = [context["""title"""] for context in dpr_record["""positive_ctxs"""]] eval_file.write(question + """\n""" ) gold_file.write("""\t""".join(lowerCAmelCase ) + """\n""" ) if __name__ == "__main__": main()
70
0
"""simple docstring""" import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) __SCREAMING_SNAKE_CASE =pytest.mark.integration @pytest.mark.parametrize('path' , ['paws', 'csv'] ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Any ): inspect_dataset(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = path + '.py' assert script_name in os.listdir(__SCREAMING_SNAKE_CASE ) assert "__pycache__" not in os.listdir(__SCREAMING_SNAKE_CASE ) @pytest.mark.filterwarnings('ignore:inspect_metric is deprecated:FutureWarning' ) @pytest.mark.filterwarnings('ignore:metric_module_factory is deprecated:FutureWarning' ) @pytest.mark.parametrize('path' , ['accuracy'] ) def lowercase__( __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Union[str, Any] ): inspect_metric(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase_ : str = path + '.py' assert script_name in os.listdir(__SCREAMING_SNAKE_CASE ) assert "__pycache__" not in os.listdir(__SCREAMING_SNAKE_CASE ) @pytest.mark.parametrize( 'path, config_name, expected_splits' , [ ('squad', 'plain_text', ['train', 'validation']), ('dalle-mini/wit', 'dalle-mini--wit', ['train']), ('paws', 'labeled_final', ['train', 'test', 'validation']), ] , ) def lowercase__( __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[int] ): lowercase_ : Tuple = get_dataset_config_info(__SCREAMING_SNAKE_CASE , config_name=__SCREAMING_SNAKE_CASE ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( 'path, config_name, expected_exception' , [ ('paws', None, ValueError), ] , ) def lowercase__( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Tuple ): with pytest.raises(__SCREAMING_SNAKE_CASE ): get_dataset_config_info(__SCREAMING_SNAKE_CASE , config_name=__SCREAMING_SNAKE_CASE ) @pytest.mark.parametrize( 'path, expected' , [ ('squad', 'plain_text'), ('acronym_identification', 'default'), ('lhoestq/squad', 'plain_text'), ('lhoestq/test', 'default'), ('lhoestq/demo1', 'lhoestq--demo1'), ('dalle-mini/wit', 'dalle-mini--wit'), ] , ) def lowercase__( __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : int ): lowercase_ : Tuple = get_dataset_config_names(__SCREAMING_SNAKE_CASE ) assert expected in config_names @pytest.mark.parametrize( 'path, expected_configs, expected_splits_in_first_config' , [ ('squad', ['plain_text'], ['train', 'validation']), ('dalle-mini/wit', ['dalle-mini--wit'], ['train']), ('paws', ['labeled_final', 'labeled_swap', 'unlabeled_final'], ['train', 'test', 'validation']), ] , ) def lowercase__( __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] ): lowercase_ : Optional[int] = get_dataset_infos(__SCREAMING_SNAKE_CASE ) assert list(infos.keys() ) == expected_configs lowercase_ : Optional[int] = expected_configs[0] assert expected_config in infos lowercase_ : Dict = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( 'path, expected_config, expected_splits' , [ ('squad', 'plain_text', ['train', 'validation']), ('dalle-mini/wit', 'dalle-mini--wit', ['train']), ('paws', 'labeled_final', ['train', 'test', 'validation']), ] , ) def lowercase__( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Dict ): lowercase_ : int = get_dataset_infos(__SCREAMING_SNAKE_CASE ) assert expected_config in infos lowercase_ : Optional[int] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( 'path, config_name, expected_exception' , [ ('paws', None, ValueError), ] , ) def lowercase__( __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[str] ): with pytest.raises(__SCREAMING_SNAKE_CASE ): get_dataset_split_names(__SCREAMING_SNAKE_CASE , config_name=__SCREAMING_SNAKE_CASE )
213
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : Optional[Any] ={ '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Union[str, Any] =[ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys A__ : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
70
0
import sys from pathlib import Path SCREAMING_SNAKE_CASE_ = Path(__file__).resolve().parents[3] / '''src''' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(4_2) SCREAMING_SNAKE_CASE_ = {'''base''': '''patrickvonplaten/wav2vec2_tiny_random''', '''robust''': '''patrickvonplaten/wav2vec2_tiny_random_robust'''} SCREAMING_SNAKE_CASE_ = '''zero2''' SCREAMING_SNAKE_CASE_ = '''zero3''' SCREAMING_SNAKE_CASE_ = [ZEROa, ZEROa] def __lowercase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = parameterized.to_safe_name("""_""".join(str(_SCREAMING_SNAKE_CASE ) for x in param.args ) ) return F"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test SCREAMING_SNAKE_CASE_ = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class UpperCamelCase__ ( snake_case_ ): '''simple docstring''' @parameterized.expand(__snake_case ,name_func=__snake_case ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[int] ) -> List[str]: '''simple docstring''' self.run_and_check( stage=__snake_case ,model=__snake_case ,distributed=__snake_case ,fpaa=__snake_case ,) @require_torch_multi_gpu @parameterized.expand(__snake_case ,name_func=__snake_case ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : int ) -> int: '''simple docstring''' self.run_and_check( stage=__snake_case ,model=__snake_case ,distributed=__snake_case ,fpaa=__snake_case ,) @parameterized.expand(__snake_case ,name_func=__snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Any ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Dict ) -> Tuple: '''simple docstring''' self.run_and_check( stage=__snake_case ,model=__snake_case ,distributed=__snake_case ,fpaa=__snake_case ,) @require_torch_multi_gpu @parameterized.expand(__snake_case ,name_func=__snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Dict ) -> Union[str, Any]: '''simple docstring''' self.run_and_check( stage=__snake_case ,model=__snake_case ,distributed=__snake_case ,fpaa=__snake_case ,) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ,lowerCamelCase__ : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE__ ( self : Dict ,lowerCamelCase__ : str ,lowerCamelCase__ : str ,lowerCamelCase__ : int = 10 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = models[model] SCREAMING_SNAKE_CASE = self.run_trainer( stage=__snake_case ,model_name=__snake_case ,eval_steps=__snake_case ,num_train_epochs=1 ,distributed=__snake_case ,fpaa=__snake_case ,) self.do_checks(__snake_case ) return output_dir def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ,lowerCamelCase__ : str ,lowerCamelCase__ : str ,lowerCamelCase__ : int = 10 ,lowerCamelCase__ : int = 1 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.get_auto_remove_tmp_dir("""./xxx""" ,after=__snake_case ) SCREAMING_SNAKE_CASE = F"""\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(__snake_case )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n """.split() if fpaa: args.extend(["""--fp16"""] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files SCREAMING_SNAKE_CASE = F"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() SCREAMING_SNAKE_CASE = [F"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] SCREAMING_SNAKE_CASE = self.get_launcher(__snake_case ) SCREAMING_SNAKE_CASE = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__snake_case ,env=self.get_env() ) return output_dir def SCREAMING_SNAKE_CASE__ ( self : Tuple ,lowerCamelCase__ : List[Any]=False ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = min(2 ,get_gpu_count() ) if distributed else 1 return F"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
296
'''simple docstring''' import math from typing import Callable, List, Optional, Union import numpy as np import PIL import torch from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from diffusers.schedulers import DDIMScheduler, DDPMScheduler, LMSDiscreteScheduler, PNDMScheduler def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=[] ): """simple docstring""" _lowerCAmelCase = size[0] - overlap_pixels * 2 _lowerCAmelCase = size[1] - overlap_pixels * 2 for letter in ["l", "r"]: if letter in remove_borders: size_x += overlap_pixels for letter in ["t", "b"]: if letter in remove_borders: size_y += overlap_pixels _lowerCAmelCase = np.ones((size_y, size_x) , dtype=np.uinta ) * 2_55 _lowerCAmelCase = np.pad(lowerCAmelCase , mode="""linear_ramp""" , pad_width=lowerCAmelCase , end_values=0 ) if "l" in remove_borders: _lowerCAmelCase = mask[:, overlap_pixels : mask.shape[1]] if "r" in remove_borders: _lowerCAmelCase = mask[:, 0 : mask.shape[1] - overlap_pixels] if "t" in remove_borders: _lowerCAmelCase = mask[overlap_pixels : mask.shape[0], :] if "b" in remove_borders: _lowerCAmelCase = mask[0 : mask.shape[0] - overlap_pixels, :] return mask def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" return max(lowerCAmelCase , min(lowerCAmelCase , lowerCAmelCase ) ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" return ( clamp(rect[0] , min[0] , max[0] ), clamp(rect[1] , min[1] , max[1] ), clamp(rect[2] , min[0] , max[0] ), clamp(rect[3] , min[1] , max[1] ), ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = list(lowerCAmelCase ) rect[0] -= overlap rect[1] -= overlap rect[2] += overlap rect[3] += overlap _lowerCAmelCase = clamp_rect(lowerCAmelCase , [0, 0] , [image_size[0], image_size[1]] ) return rect def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = Image.new("""RGB""" , (tile.size[0] + original_slice, tile.size[1]) ) result.paste( original_image.resize((tile.size[0], tile.size[1]) , Image.BICUBIC ).crop( (slice_x, 0, slice_x + original_slice, tile.size[1]) ) , (0, 0) , ) result.paste(lowerCAmelCase , (original_slice, 0) ) return result def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = (original_image_slice * 4, 0, tile.size[0], tile.size[1]) _lowerCAmelCase = tile.crop(lowerCAmelCase ) return tile def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = n % d return n - divisor class UpperCAmelCase ( snake_case_ ): def __init__( self : List[Any] , __snake_case : AutoencoderKL , __snake_case : CLIPTextModel , __snake_case : CLIPTokenizer , __snake_case : UNetaDConditionModel , __snake_case : DDPMScheduler , __snake_case : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , __snake_case : int = 3_50 , ) -> int: super().__init__( vae=__snake_case , text_encoder=__snake_case , tokenizer=__snake_case , unet=__snake_case , low_res_scheduler=__snake_case , scheduler=__snake_case , max_noise_level=__snake_case , ) def lowercase__ ( self : List[Any] , __snake_case : List[Any] , __snake_case : List[Any] , __snake_case : Dict , __snake_case : List[Any] , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Any , **__snake_case : str ) -> int: torch.manual_seed(0 ) _lowerCAmelCase = ( min(image.size[0] - (tile_size + original_image_slice) , x * tile_size ), min(image.size[1] - (tile_size + original_image_slice) , y * tile_size ), min(image.size[0] , (x + 1) * tile_size ), min(image.size[1] , (y + 1) * tile_size ), ) _lowerCAmelCase = add_overlap_rect(__snake_case , __snake_case , image.size ) _lowerCAmelCase = image.crop(__snake_case ) _lowerCAmelCase = ((crop_rect[0] + ((crop_rect[2] - crop_rect[0]) / 2)) / image.size[0]) * tile.size[0] _lowerCAmelCase = translated_slice_x - (original_image_slice / 2) _lowerCAmelCase = max(0 , __snake_case ) _lowerCAmelCase = squeeze_tile(__snake_case , __snake_case , __snake_case , __snake_case ) _lowerCAmelCase = to_input.size _lowerCAmelCase = to_input.resize((tile_size, tile_size) , Image.BICUBIC ) _lowerCAmelCase = super(__snake_case , self ).__call__(image=__snake_case , **__snake_case ).images[0] _lowerCAmelCase = upscaled_tile.resize((orig_input_size[0] * 4, orig_input_size[1] * 4) , Image.BICUBIC ) _lowerCAmelCase = unsqueeze_tile(__snake_case , __snake_case ) _lowerCAmelCase = upscaled_tile.resize((tile.size[0] * 4, tile.size[1] * 4) , Image.BICUBIC ) _lowerCAmelCase = [] if x == 0: remove_borders.append("""l""" ) elif crop_rect[2] == image.size[0]: remove_borders.append("""r""" ) if y == 0: remove_borders.append("""t""" ) elif crop_rect[3] == image.size[1]: remove_borders.append("""b""" ) _lowerCAmelCase = Image.fromarray( make_transparency_mask( (upscaled_tile.size[0], upscaled_tile.size[1]) , tile_border * 4 , remove_borders=__snake_case ) , mode="""L""" , ) final_image.paste( __snake_case , (crop_rect_with_overlap[0] * 4, crop_rect_with_overlap[1] * 4) , __snake_case ) @torch.no_grad() def __call__( self : Union[str, Any] , __snake_case : Union[str, List[str]] , __snake_case : Union[PIL.Image.Image, List[PIL.Image.Image]] , __snake_case : int = 75 , __snake_case : float = 9.0 , __snake_case : int = 50 , __snake_case : Optional[Union[str, List[str]]] = None , __snake_case : Optional[int] = 1 , __snake_case : float = 0.0 , __snake_case : Optional[torch.Generator] = None , __snake_case : Optional[torch.FloatTensor] = None , __snake_case : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __snake_case : int = 1 , __snake_case : int = 1_28 , __snake_case : int = 32 , __snake_case : int = 32 , ) -> str: _lowerCAmelCase = Image.new("""RGB""" , (image.size[0] * 4, image.size[1] * 4) ) _lowerCAmelCase = math.ceil(image.size[0] / tile_size ) _lowerCAmelCase = math.ceil(image.size[1] / tile_size ) _lowerCAmelCase = tcx * tcy _lowerCAmelCase = 0 for y in range(__snake_case ): for x in range(__snake_case ): self._process_tile( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , prompt=__snake_case , num_inference_steps=__snake_case , guidance_scale=__snake_case , noise_level=__snake_case , negative_prompt=__snake_case , num_images_per_prompt=__snake_case , eta=__snake_case , generator=__snake_case , latents=__snake_case , ) current_count += 1 if callback is not None: callback({"""progress""": current_count / total_tile_count, """image""": final_image} ) return final_image def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = """stabilityai/stable-diffusion-x4-upscaler""" _lowerCAmelCase = StableDiffusionTiledUpscalePipeline.from_pretrained(lowerCAmelCase , revision="""fp16""" , torch_dtype=torch.floataa ) _lowerCAmelCase = pipe.to("""cuda""" ) _lowerCAmelCase = Image.open("""../../docs/source/imgs/diffusers_library.jpg""" ) def callback(lowerCAmelCase ): print(f"progress: {obj['progress']:.4f}" ) obj["image"].save("""diffusers_library_progress.jpg""" ) _lowerCAmelCase = pipe(image=lowerCAmelCase , prompt="""Black font, white background, vector""" , noise_level=40 , callback=lowerCAmelCase ) final_image.save("""diffusers_library.jpg""" ) if __name__ == "__main__": main()
70
0
"""simple docstring""" from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase ( snake_case_ ): '''simple docstring''' def _a (self ): """simple docstring""" UpperCAmelCase__ : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__snake_case , """embed_dim""" ) ) self.parent.assertTrue(hasattr(__snake_case , """num_heads""" ) ) class lowerCamelCase : '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=64 , _lowerCamelCase=3 , _lowerCamelCase=[16, 48, 96] , _lowerCamelCase=[1, 3, 6] , _lowerCamelCase=[1, 2, 10] , _lowerCamelCase=[7, 3, 3] , _lowerCamelCase=[4, 2, 2] , _lowerCamelCase=[2, 1, 1] , _lowerCamelCase=[2, 2, 2] , _lowerCamelCase=[False, False, True] , _lowerCamelCase=[0.0, 0.0, 0.0] , _lowerCamelCase=0.02 , _lowerCamelCase=1e-12 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=2 , ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : Any = image_size UpperCAmelCase__ : int = patch_sizes UpperCAmelCase__ : Dict = patch_stride UpperCAmelCase__ : int = patch_padding UpperCAmelCase__ : Optional[Any] = is_training UpperCAmelCase__ : List[str] = use_labels UpperCAmelCase__ : int = num_labels UpperCAmelCase__ : List[Any] = num_channels UpperCAmelCase__ : List[str] = embed_dim UpperCAmelCase__ : Optional[Any] = num_heads UpperCAmelCase__ : Tuple = stride_kv UpperCAmelCase__ : Tuple = depth UpperCAmelCase__ : Tuple = cls_token UpperCAmelCase__ : int = attention_drop_rate UpperCAmelCase__ : List[Any] = initializer_range UpperCAmelCase__ : int = layer_norm_eps def _a (self ): """simple docstring""" UpperCAmelCase__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : int = None if self.use_labels: # create a random int32 tensor of given shape UpperCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase__ : List[Any] = self.get_config() return config, pixel_values, labels def _a (self ): """simple docstring""" return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Dict = TFCvtModel(config=__snake_case ) UpperCAmelCase__ : Union[str, Any] = model(__snake_case , training=__snake_case ) UpperCAmelCase__ : Any = (self.image_size, self.image_size) UpperCAmelCase__ , UpperCAmelCase__ : str = image_size[0], image_size[1] for i in range(len(self.depth ) ): UpperCAmelCase__ : Any = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) UpperCAmelCase__ : int = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : int = self.num_labels UpperCAmelCase__ : Any = TFCvtForImageClassification(__snake_case ) UpperCAmelCase__ : Tuple = model(__snake_case , labels=__snake_case , training=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = config_and_inputs UpperCAmelCase__ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class lowerCamelCase ( snake_case_ , snake_case_ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () SCREAMING_SNAKE_CASE = ( {'''feature-extraction''': TFCvtModel, '''image-classification''': TFCvtForImageClassification} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = TFCvtModelTester(self ) UpperCAmelCase__ : Optional[int] = TFCvtConfigTester(self , config_class=__snake_case , has_text_modality=__snake_case , hidden_size=37 ) def _a (self ): """simple docstring""" self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason="""Cvt does not output attentions""" ) def _a (self ): """simple docstring""" pass @unittest.skip(reason="""Cvt does not use inputs_embeds""" ) def _a (self ): """simple docstring""" pass @unittest.skip(reason="""Cvt does not support input and output embeddings""" ) def _a (self ): """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) def _a (self ): """simple docstring""" super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def _a (self ): """simple docstring""" super().test_keras_fit() @unittest.skip(reason="""Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8""" ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = tf.keras.mixed_precision.Policy("""mixed_float16""" ) tf.keras.mixed_precision.set_global_policy(__snake_case ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy("""float32""" ) def _a (self ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Union[str, Any] = model_class(__snake_case ) UpperCAmelCase__ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : str = [*signature.parameters.keys()] UpperCAmelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __snake_case ) def _a (self ): """simple docstring""" def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): UpperCAmelCase__ : int = model_class(__snake_case ) UpperCAmelCase__ : Tuple = model(**self._prepare_for_class(__snake_case , __snake_case ) ) UpperCAmelCase__ : Dict = outputs.hidden_states UpperCAmelCase__ : Optional[int] = len(self.model_tester.depth ) self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase__ : Union[str, Any] = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def _a (self ): """simple docstring""" UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__snake_case ) @slow def _a (self ): """simple docstring""" for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : Tuple = TFCvtModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def a__ ( ) -> Union[str, Any]: UpperCAmelCase__ : Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def _a (self ): """simple docstring""" return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCAmelCase__ : Any = self.default_image_processor UpperCAmelCase__ : Optional[int] = prepare_img() UpperCAmelCase__ : List[Any] = image_processor(images=__snake_case , return_tensors="""tf""" ) # forward pass UpperCAmelCase__ : str = model(**__snake_case ) # verify the logits UpperCAmelCase__ : Dict = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , __snake_case ) UpperCAmelCase__ : Optional[Any] = tf.constant([0.9_285, 0.9_015, -0.3_150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , __snake_case , atol=1e-4 ) )
171
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class UpperCAmelCase ( snake_case_ , unittest.TestCase ): _lowercase: int = KandinskyVaaImgaImgPipeline _lowercase: List[str] = ['''image_embeds''', '''negative_image_embeds''', '''image'''] _lowercase: Optional[int] = [ '''image_embeds''', '''negative_image_embeds''', '''image''', ] _lowercase: Tuple = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] _lowercase: List[str] = False @property def lowercase__ ( self : str ) -> List[str]: return 32 @property def lowercase__ ( self : Optional[int] ) -> List[Any]: return 32 @property def lowercase__ ( self : Tuple ) -> str: return self.time_input_dim @property def lowercase__ ( self : Any ) -> Optional[int]: return self.time_input_dim * 4 @property def lowercase__ ( self : int ) -> Optional[Any]: return 1_00 @property def lowercase__ ( self : int ) -> Dict: torch.manual_seed(0 ) _lowerCAmelCase = { """in_channels""": 4, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } _lowerCAmelCase = UNetaDConditionModel(**__snake_case ) return model @property def lowercase__ ( self : Union[str, Any] ) -> Tuple: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowercase__ ( self : Dict ) -> str: torch.manual_seed(0 ) _lowerCAmelCase = VQModel(**self.dummy_movq_kwargs ) return model def lowercase__ ( self : Optional[int] ) -> Optional[int]: _lowerCAmelCase = self.dummy_unet _lowerCAmelCase = self.dummy_movq _lowerCAmelCase = { """num_train_timesteps""": 10_00, """beta_schedule""": """linear""", """beta_start""": 0.0_00_85, """beta_end""": 0.0_12, """clip_sample""": False, """set_alpha_to_one""": False, """steps_offset""": 0, """prediction_type""": """epsilon""", """thresholding""": False, } _lowerCAmelCase = DDIMScheduler(**__snake_case ) _lowerCAmelCase = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def lowercase__ ( self : int , __snake_case : List[str] , __snake_case : List[Any]=0 ) -> Union[str, Any]: _lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __snake_case ) # create init_image _lowerCAmelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((2_56, 2_56) ) if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = { """image""": init_image, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 10, """guidance_scale""": 7.0, """strength""": 0.2, """output_type""": """np""", } return inputs def lowercase__ ( self : str ) -> Tuple: _lowerCAmelCase = """cpu""" _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) _lowerCAmelCase = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = pipe(**self.get_dummy_inputs(__snake_case ) ) _lowerCAmelCase = output.images _lowerCAmelCase = pipe( **self.get_dummy_inputs(__snake_case ) , return_dict=__snake_case , )[0] _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _lowerCAmelCase = np.array( [0.6_19_97_78, 0.63_98_44_06, 0.46_14_57_85, 0.62_94_49_84, 0.5_62_22_15, 0.47_30_61_32, 0.47_44_14_56, 0.4_60_76_06, 0.48_71_92_63] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): def lowercase__ ( self : Any ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : int ) -> Dict: _lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_img2img_frog.npy""" ) _lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) _lowerCAmelCase = """A red cartoon frog, 4k""" _lowerCAmelCase = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__snake_case ) _lowerCAmelCase = KandinskyVaaImgaImgPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder""" , torch_dtype=torch.floataa ) _lowerCAmelCase = pipeline.to(__snake_case ) pipeline.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase , _lowerCAmelCase = pipe_prior( __snake_case , generator=__snake_case , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() _lowerCAmelCase = pipeline( image=__snake_case , image_embeds=__snake_case , negative_image_embeds=__snake_case , generator=__snake_case , num_inference_steps=1_00 , height=7_68 , width=7_68 , strength=0.2 , output_type="""np""" , ) _lowerCAmelCase = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(__snake_case , __snake_case )
70
0
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput _UpperCamelCase = '''scheduler_config.json''' class __lowercase (snake_case_ ): _UpperCamelCase = 1 _UpperCamelCase = 2 _UpperCamelCase = 3 _UpperCamelCase = 4 _UpperCamelCase = 5 @dataclass class __lowercase (snake_case_ ): _UpperCamelCase = 42 class __lowercase : _UpperCamelCase = SCHEDULER_CONFIG_NAME _UpperCamelCase = ['''dtype'''] _UpperCamelCase = [] _UpperCamelCase = True @classmethod def UpperCamelCase__ ( cls , A_ = None , A_ = None , A_=False , **A_ , ) ->int: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : int = cls.load_config( pretrained_model_name_or_path=__snake_case , subfolder=__snake_case , return_unused_kwargs=__snake_case , **__snake_case , ) __lowerCAmelCase, __lowerCAmelCase : int = cls.from_config(__snake_case , return_unused_kwargs=__snake_case , **__snake_case ) if hasattr(__snake_case , '''create_state''' ) and getattr(__snake_case , '''has_state''' , __snake_case ): __lowerCAmelCase : Any = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCamelCase__ ( self , A_ , A_ = False , **A_ ) ->int: '''simple docstring''' self.save_config(save_directory=__snake_case , push_to_hub=__snake_case , **__snake_case ) @property def UpperCamelCase__ ( self ) ->Dict: '''simple docstring''' return self._get_compatibles() @classmethod def UpperCamelCase__ ( cls ) ->Optional[int]: '''simple docstring''' __lowerCAmelCase : Dict = list(set([cls.__name__] + cls._compatibles ) ) __lowerCAmelCase : Optional[int] = importlib.import_module(__name__.split('''.''' )[0] ) __lowerCAmelCase : Tuple = [ getattr(__snake_case , __snake_case ) for c in compatible_classes_str if hasattr(__snake_case , __snake_case ) ] return compatible_classes def _lowercase ( lowercase__ , lowercase__ ): assert len(lowercase__ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(lowercase__ ) - x.ndim) ) , lowercase__ ) def _lowercase ( lowercase__ , lowercase__=0.9_9_9 , lowercase__=jnp.floataa ): def alpha_bar(lowercase__ ): return math.cos((time_step + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 __lowerCAmelCase : List[str] = [] for i in range(lowercase__ ): __lowerCAmelCase : str = i / num_diffusion_timesteps __lowerCAmelCase : List[Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(lowercase__ ) / alpha_bar(lowercase__ ) , lowercase__ ) ) return jnp.array(lowercase__ , dtype=lowercase__ ) @flax.struct.dataclass class __lowercase : _UpperCamelCase = 42 _UpperCamelCase = 42 _UpperCamelCase = 42 @classmethod def UpperCamelCase__ ( cls , A_ ) ->int: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = scheduler.config if config.trained_betas is not None: __lowerCAmelCase : Union[str, Any] = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": __lowerCAmelCase : Union[str, Any] = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. __lowerCAmelCase : Any = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule __lowerCAmelCase : List[str] = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( f"""beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}""" ) __lowerCAmelCase : str = 1.0 - betas __lowerCAmelCase : Union[str, Any] = jnp.cumprod(__snake_case , axis=0 ) return cls( alphas=__snake_case , betas=__snake_case , alphas_cumprod=__snake_case , ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase : Optional[Any] = state.alphas_cumprod __lowerCAmelCase : Any = alphas_cumprod[timesteps] ** 0.5 __lowerCAmelCase : int = sqrt_alpha_prod.flatten() __lowerCAmelCase : Tuple = broadcast_to_shape_from_left(lowercase__ , original_samples.shape ) __lowerCAmelCase : Optional[int] = (1 - alphas_cumprod[timesteps]) ** 0.5 __lowerCAmelCase : Any = sqrt_one_minus_alpha_prod.flatten() __lowerCAmelCase : str = broadcast_to_shape_from_left(lowercase__ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase, __lowerCAmelCase : str = get_sqrt_alpha_prod(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowerCAmelCase : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): __lowerCAmelCase, __lowerCAmelCase : List[str] = get_sqrt_alpha_prod(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) __lowerCAmelCase : Union[str, Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
275
'''simple docstring''' import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class UpperCAmelCase ( snake_case_ ): def lowercase__ ( self : List[Any] ) -> Union[str, Any]: _lowerCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__snake_case , """hidden_sizes""" ) ) self.parent.assertTrue(hasattr(__snake_case , """num_attention_heads""" ) ) self.parent.assertTrue(hasattr(__snake_case , """num_encoder_blocks""" ) ) class UpperCAmelCase : def __init__( self : Optional[int] , __snake_case : str , __snake_case : Dict=13 , __snake_case : str=64 , __snake_case : Dict=3 , __snake_case : Dict=4 , __snake_case : Tuple=[2, 2, 2, 2] , __snake_case : int=[8, 4, 2, 1] , __snake_case : List[str]=[16, 32, 64, 1_28] , __snake_case : Optional[Any]=[1, 4, 8, 16] , __snake_case : Dict=[1, 2, 4, 8] , __snake_case : Optional[Any]=True , __snake_case : List[str]=True , __snake_case : int="gelu" , __snake_case : Optional[Any]=0.1 , __snake_case : Any=0.1 , __snake_case : Tuple=0.02 , __snake_case : Union[str, Any]=3 , __snake_case : Tuple=None , ) -> List[str]: _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = image_size _lowerCAmelCase = num_channels _lowerCAmelCase = num_encoder_blocks _lowerCAmelCase = sr_ratios _lowerCAmelCase = depths _lowerCAmelCase = hidden_sizes _lowerCAmelCase = downsampling_rates _lowerCAmelCase = num_attention_heads _lowerCAmelCase = is_training _lowerCAmelCase = use_labels _lowerCAmelCase = hidden_act _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = initializer_range _lowerCAmelCase = num_labels _lowerCAmelCase = scope def lowercase__ ( self : int ) -> Union[str, Any]: _lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase = None if self.use_labels: _lowerCAmelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCAmelCase = self.get_config() return config, pixel_values, labels def lowercase__ ( self : List[Any] ) -> List[str]: return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def lowercase__ ( self : Tuple , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : Optional[int] ) -> Tuple: _lowerCAmelCase = SegformerModel(config=__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = _lowerCAmelCase = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def lowercase__ ( self : List[str] , __snake_case : List[Any] , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> List[str]: _lowerCAmelCase = self.num_labels _lowerCAmelCase = SegformerForSemanticSegmentation(__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = model(__snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) _lowerCAmelCase = model(__snake_case , labels=__snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def lowercase__ ( self : str , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Dict ) -> List[str]: _lowerCAmelCase = 1 _lowerCAmelCase = SegformerForSemanticSegmentation(config=__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(__snake_case ) _lowerCAmelCase = model(__snake_case , labels=__snake_case ) self.parent.assertGreater(result.loss , 0.0 ) def lowercase__ ( self : Optional[int] ) -> int: _lowerCAmelCase = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = config_and_inputs _lowerCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Any = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) _lowercase: Tuple = ( { '''feature-extraction''': SegformerModel, '''image-classification''': SegformerForImageClassification, '''image-segmentation''': SegformerForSemanticSegmentation, } if is_torch_available() else {} ) _lowercase: Tuple = True _lowercase: Union[str, Any] = False _lowercase: Dict = False _lowercase: Optional[Any] = False def lowercase__ ( self : Tuple ) -> Any: _lowerCAmelCase = SegformerModelTester(self ) _lowerCAmelCase = SegformerConfigTester(self , config_class=__snake_case ) def lowercase__ ( self : Optional[Any] ) -> Dict: self.config_tester.run_common_tests() def lowercase__ ( self : int ) -> Union[str, Any]: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def lowercase__ ( self : Dict ) -> int: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*__snake_case ) def lowercase__ ( self : Dict ) -> Dict: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*__snake_case ) @unittest.skip("""SegFormer does not use inputs_embeds""" ) def lowercase__ ( self : int ) -> Union[str, Any]: pass @unittest.skip("""SegFormer does not have get_input_embeddings method and get_output_embeddings methods""" ) def lowercase__ ( self : Optional[int] ) -> int: pass def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase = model_class(__snake_case ) _lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase = [*signature.parameters.keys()] _lowerCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __snake_case ) def lowercase__ ( self : Tuple ) -> Tuple: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase = True for model_class in self.all_model_classes: _lowerCAmelCase = True _lowerCAmelCase = False _lowerCAmelCase = True _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) _lowerCAmelCase = outputs.attentions _lowerCAmelCase = sum(self.model_tester.depths ) self.assertEqual(len(__snake_case ) , __snake_case ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _lowerCAmelCase = True _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) _lowerCAmelCase = outputs.attentions self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first attentions (first block, first layer) _lowerCAmelCase = (self.model_tester.image_size // 4) ** 2 _lowerCAmelCase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) _lowerCAmelCase = (self.model_tester.image_size // 32) ** 2 _lowerCAmelCase = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) _lowerCAmelCase = len(__snake_case ) # Check attention is always last and order is fine _lowerCAmelCase = True _lowerCAmelCase = True _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) self.assertEqual(out_len + 1 , len(__snake_case ) ) _lowerCAmelCase = outputs.attentions self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first attentions (first block, first layer) _lowerCAmelCase = (self.model_tester.image_size // 4) ** 2 _lowerCAmelCase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def lowercase__ ( self : int ) -> List[str]: def check_hidden_states_output(__snake_case : str , __snake_case : Tuple , __snake_case : Optional[int] ): _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) _lowerCAmelCase = outputs.hidden_states _lowerCAmelCase = self.model_tester.num_encoder_blocks self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCAmelCase = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) def lowercase__ ( self : Optional[Any] ) -> Any: if not self.model_tester.is_training: return _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase = True for model_class in self.all_model_classes: if model_class in get_values(__snake_case ): continue _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.train() _lowerCAmelCase = self._prepare_for_class(__snake_case , __snake_case , return_labels=__snake_case ) _lowerCAmelCase = model(**__snake_case ).loss loss.backward() @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase__ ( self : Tuple ) -> Dict: pass @slow def lowercase__ ( self : str ) -> Optional[int]: for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase = SegformerModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class UpperCAmelCase ( unittest.TestCase ): @slow def lowercase__ ( self : Union[str, Any] ) -> Any: # only resize + normalize _lowerCAmelCase = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) _lowerCAmelCase = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( __snake_case ) _lowerCAmelCase = prepare_img() _lowerCAmelCase = image_processor(images=__snake_case , return_tensors="""pt""" ) _lowerCAmelCase = encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = torch.Size((1, model.config.num_labels, 1_28, 1_28) ) self.assertEqual(outputs.logits.shape , __snake_case ) _lowerCAmelCase = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def lowercase__ ( self : Optional[Any] ) -> Any: # only resize + normalize _lowerCAmelCase = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) _lowerCAmelCase = SegformerForSemanticSegmentation.from_pretrained( """nvidia/segformer-b1-finetuned-cityscapes-1024-1024""" ).to(__snake_case ) _lowerCAmelCase = prepare_img() _lowerCAmelCase = image_processor(images=__snake_case , return_tensors="""pt""" ) _lowerCAmelCase = encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = torch.Size((1, model.config.num_labels, 1_28, 1_28) ) self.assertEqual(outputs.logits.shape , __snake_case ) _lowerCAmelCase = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , __snake_case , atol=1E-1 ) ) @slow def lowercase__ ( self : Any ) -> str: # only resize + normalize _lowerCAmelCase = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) _lowerCAmelCase = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( __snake_case ) _lowerCAmelCase = prepare_img() _lowerCAmelCase = image_processor(images=__snake_case , return_tensors="""pt""" ) _lowerCAmelCase = encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = outputs.logits.detach().cpu() _lowerCAmelCase = image_processor.post_process_semantic_segmentation(outputs=__snake_case , target_sizes=[(5_00, 3_00)] ) _lowerCAmelCase = torch.Size((5_00, 3_00) ) self.assertEqual(segmentation[0].shape , __snake_case ) _lowerCAmelCase = image_processor.post_process_semantic_segmentation(outputs=__snake_case ) _lowerCAmelCase = torch.Size((1_28, 1_28) ) self.assertEqual(segmentation[0].shape , __snake_case )
70
0
'''simple docstring''' lowerCamelCase_ = '''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip lowerCamelCase_ = concatenate_datasets lowerCamelCase_ = DownloadConfig lowerCamelCase_ = DownloadManager lowerCamelCase_ = DownloadMode lowerCamelCase_ = DownloadConfig lowerCamelCase_ = DownloadMode lowerCamelCase_ = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
79
'''simple docstring''' from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class UpperCAmelCase : _lowercase: List[str] _lowercase: Optional[str] = None # Automatically constructed _lowercase: ClassVar[str] = "dict" _lowercase: ClassVar[Any] = None _lowercase: str = field(default='''Translation''' , init=snake_case_ , repr=snake_case_ ) def __call__( self : Optional[int] ) -> Optional[int]: return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def lowercase__ ( self : Union[str, Any] ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value("""string""" ) for k in sorted(self.languages )} @dataclass class UpperCAmelCase : _lowercase: Optional[List] = None _lowercase: Optional[int] = None _lowercase: Optional[str] = None # Automatically constructed _lowercase: ClassVar[str] = "dict" _lowercase: ClassVar[Any] = None _lowercase: str = field(default='''TranslationVariableLanguages''' , init=snake_case_ , repr=snake_case_ ) def lowercase__ ( self : Any ) -> Optional[Any]: _lowerCAmelCase = sorted(set(self.languages ) ) if self.languages else None _lowerCAmelCase = len(self.languages ) if self.languages else None def __call__( self : List[str] ) -> Optional[Any]: return pa.struct({"""language""": pa.list_(pa.string() ), """translation""": pa.list_(pa.string() )} ) def lowercase__ ( self : Optional[Any] , __snake_case : Tuple ) -> Any: _lowerCAmelCase = set(self.languages ) if self.languages and set(__snake_case ) - lang_set: raise ValueError( f"Some languages in example ({', '.join(sorted(set(__snake_case ) - lang_set ) )}) are not in valid set ({', '.join(__snake_case )})." ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. _lowerCAmelCase = [] for lang, text in translation_dict.items(): if isinstance(__snake_case , __snake_case ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. _lowerCAmelCase , _lowerCAmelCase = zip(*sorted(__snake_case ) ) return {"language": languages, "translation": translations} def lowercase__ ( self : str ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value("""string""" ) ), "translation": Sequence(Value("""string""" ) ), }
70
0
def __a ( SCREAMING_SNAKE_CASE ) -> Optional[int]: '''simple docstring''' if upper_limit < 0: raise ValueError('''Limit for the Catalan sequence must be ≥ 0''' ) __UpperCAmelCase = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 __UpperCAmelCase = 1 if upper_limit > 0: __UpperCAmelCase = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(SCREAMING_SNAKE_CASE ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print('\n********* Catalan Numbers Using Dynamic Programming ************\n') print('\n*** Enter -1 at any time to quit ***') print('\nEnter the upper limit (≥ 0) for the Catalan number sequence: ', end='') try: while True: A_ : Tuple = int(input().strip()) if N < 0: print('\n********* Goodbye!! ************') break else: print(F"""The Catalan numbers from 0 through {N} are:""") print(catalan_numbers(N)) print('Try another upper limit for the sequence: ', end='') except (NameError, ValueError): print('\n********* Invalid input, goodbye! ************\n') import doctest doctest.testmod()
333
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer A__ : List[str] =logging.get_logger(__name__) A__ : Any ={'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A__ : Any ={ '''vocab_file''': { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt''' ), } } A__ : Optional[int] ={ '''junnyu/roformer_chinese_small''': 15_36, '''junnyu/roformer_chinese_base''': 15_36, '''junnyu/roformer_chinese_char_small''': 5_12, '''junnyu/roformer_chinese_char_base''': 5_12, '''junnyu/roformer_small_discriminator''': 1_28, '''junnyu/roformer_small_generator''': 1_28, } A__ : Optional[int] ={ '''junnyu/roformer_chinese_small''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_base''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_char_small''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_char_base''': {'''do_lower_case''': True}, '''junnyu/roformer_small_discriminator''': {'''do_lower_case''': True}, '''junnyu/roformer_small_generator''': {'''do_lower_case''': True}, } class UpperCAmelCase ( snake_case_ ): _lowercase: Optional[Any] = VOCAB_FILES_NAMES _lowercase: Tuple = PRETRAINED_VOCAB_FILES_MAP _lowercase: Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase: str = PRETRAINED_INIT_CONFIGURATION _lowercase: List[Any] = RoFormerTokenizer def __init__( self : Dict , __snake_case : str=None , __snake_case : Tuple=None , __snake_case : List[Any]=True , __snake_case : str="[UNK]" , __snake_case : Tuple="[SEP]" , __snake_case : str="[PAD]" , __snake_case : str="[CLS]" , __snake_case : Any="[MASK]" , __snake_case : Dict=True , __snake_case : str=None , **__snake_case : Optional[Any] , ) -> Union[str, Any]: super().__init__( __snake_case , tokenizer_file=__snake_case , do_lower_case=__snake_case , unk_token=__snake_case , sep_token=__snake_case , pad_token=__snake_case , cls_token=__snake_case , mask_token=__snake_case , tokenize_chinese_chars=__snake_case , strip_accents=__snake_case , **__snake_case , ) _lowerCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get("""lowercase""" , __snake_case ) != do_lower_case or pre_tok_state.get("""strip_accents""" , __snake_case ) != strip_accents ): _lowerCAmelCase = getattr(__snake_case , pre_tok_state.pop("""type""" ) ) _lowerCAmelCase = do_lower_case _lowerCAmelCase = strip_accents _lowerCAmelCase = pre_tok_class(**__snake_case ) _lowerCAmelCase = do_lower_case def __getstate__( self : int ) -> Optional[int]: _lowerCAmelCase = self.__dict__.copy() _lowerCAmelCase = BertPreTokenizer() return state def __setstate__( self : Tuple , __snake_case : Tuple ) -> List[str]: _lowerCAmelCase = d _lowerCAmelCase = self.__dict__["""_tokenizer"""].get_vocab() _lowerCAmelCase = PreTokenizer.custom(JiebaPreTokenizer(__snake_case ) ) def lowercase__ ( self : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[int]=None ) -> Optional[Any]: _lowerCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase__ ( self : List[str] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: _lowerCAmelCase = [self.sep_token_id] _lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase__ ( self : int , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: _lowerCAmelCase = self._tokenizer.model.save(__snake_case , name=__snake_case ) return tuple(__snake_case ) def lowercase__ ( self : Dict , __snake_case : Dict , __snake_case : int=None , __snake_case : List[Any]=None , __snake_case : List[Any]=False , **__snake_case : Dict , ) -> str: _lowerCAmelCase = BertPreTokenizer() return super().save_pretrained(__snake_case , __snake_case , __snake_case , __snake_case , **__snake_case )
70
0
import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class A_ (snake_case_ ): def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = tempfile.mkdtemp() UpperCAmelCase = 5 # Realm tok UpperCAmelCase = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''test''', '''question''', '''this''', '''is''', '''the''', '''first''', '''second''', '''third''', '''fourth''', '''fifth''', '''record''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] UpperCAmelCase = os.path.join(self.tmpdirname , '''realm_tokenizer''' ) os.makedirs(__snake_case , exist_ok=__snake_case ) UpperCAmelCase = os.path.join(__snake_case , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) UpperCAmelCase = os.path.join(self.tmpdirname , '''realm_block_records''' ) os.makedirs(__snake_case , exist_ok=__snake_case ) def _lowercase ( self ): '''simple docstring''' return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname , '''realm_tokenizer''' ) ) def _lowercase ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = RealmConfig(num_block_records=self.num_block_records ) return config def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = Dataset.from_dict( { '''id''': ['''0''', '''1'''], '''question''': ['''foo''', '''bar'''], '''answers''': [['''Foo''', '''Bar'''], ['''Bar''']], } ) return dataset def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = np.array( [ b'''This is the first record''', b'''This is the second record''', b'''This is the third record''', b'''This is the fourth record''', b'''This is the fifth record''', b'''This is a longer longer longer record''', ] , dtype=__snake_case , ) return block_records def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = RealmRetriever( block_records=self.get_dummy_block_records() , tokenizer=self.get_tokenizer() , ) return retriever def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.get_config() UpperCAmelCase = self.get_dummy_retriever() UpperCAmelCase = retriever.tokenizer UpperCAmelCase = np.array([0, 3] , dtype='''long''' ) UpperCAmelCase = tokenizer(['''Test question'''] ).input_ids UpperCAmelCase = tokenizer( ['''the fourth'''] , add_special_tokens=__snake_case , return_token_type_ids=__snake_case , return_attention_mask=__snake_case , ).input_ids UpperCAmelCase = config.reader_seq_len UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = retriever( __snake_case , __snake_case , answer_ids=__snake_case , max_length=__snake_case , return_tensors='''np''' ) self.assertEqual(len(__snake_case ) , 2 ) self.assertEqual(len(__snake_case ) , 2 ) self.assertEqual(len(__snake_case ) , 2 ) self.assertEqual(concat_inputs.input_ids.shape , (2, 1_0) ) self.assertEqual(concat_inputs.attention_mask.shape , (2, 1_0) ) self.assertEqual(concat_inputs.token_type_ids.shape , (2, 1_0) ) self.assertEqual(concat_inputs.special_tokens_mask.shape , (2, 1_0) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''first''', '''record''', '''[SEP]'''] , ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) , ['''[CLS]''', '''test''', '''question''', '''[SEP]''', '''this''', '''is''', '''the''', '''fourth''', '''record''', '''[SEP]'''] , ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.get_config() UpperCAmelCase = self.get_dummy_retriever() UpperCAmelCase = retriever.tokenizer UpperCAmelCase = np.array([0, 3, 5] , dtype='''long''' ) UpperCAmelCase = tokenizer(['''Test question'''] ).input_ids UpperCAmelCase = tokenizer( ['''the fourth''', '''longer longer'''] , add_special_tokens=__snake_case , return_token_type_ids=__snake_case , return_attention_mask=__snake_case , ).input_ids UpperCAmelCase = config.reader_seq_len UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = retriever( __snake_case , __snake_case , answer_ids=__snake_case , max_length=__snake_case , return_tensors='''np''' ) self.assertEqual([False, True, True] , __snake_case ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] , __snake_case ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] , __snake_case ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) # Test local path UpperCAmelCase = retriever.from_pretrained(os.path.join(self.tmpdirname , '''realm_block_records''' ) ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' ) # Test mocked remote path with patch('''transformers.models.realm.retrieval_realm.hf_hub_download''' ) as mock_hf_hub_download: UpperCAmelCase = os.path.join( os.path.join(self.tmpdirname , '''realm_block_records''' ) , _REALM_BLOCK_RECORDS_FILENAME ) UpperCAmelCase = RealmRetriever.from_pretrained('''google/realm-cc-news-pretrained-openqa''' ) self.assertEqual(retriever.block_records[0] , b'''This is the first record''' )
273
'''simple docstring''' # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class UpperCAmelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Optional[int] = StableDiffusionControlNetImgaImgPipeline _lowercase: Tuple = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Tuple = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'''control_image'''} ) _lowercase: Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase__ ( self : List[str] ) -> List[str]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Any , __snake_case : str , __snake_case : Any=0 ) -> str: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ) _lowerCAmelCase = floats_tensor(control_image.shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : Optional[int] ) -> List[Any]: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Any = StableDiffusionControlNetImgaImgPipeline _lowercase: Dict = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Any = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(__snake_case : Optional[Any] ): if isinstance(__snake_case , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = MultiControlNetModel([controlneta, controlneta] ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Tuple , __snake_case : int , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), ] _lowerCAmelCase = floats_tensor(control_image[0].shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : List[str] ) -> Dict: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) _lowerCAmelCase = 10.0 _lowerCAmelCase = 4 _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 def lowercase__ ( self : int ) -> str: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Optional[Any] ) -> Dict: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : int ) -> str: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(__snake_case ) except NotImplementedError: pass @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): def lowercase__ ( self : Union[str, Any] ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : List[str] ) -> Any: _lowerCAmelCase = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) _lowerCAmelCase = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=__snake_case , controlnet=__snake_case ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase = """evil space-punk bird""" _lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = pipe( __snake_case , __snake_case , control_image=__snake_case , generator=__snake_case , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) _lowerCAmelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) _lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9E-2
70
0
'''simple docstring''' import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase_ ( snake_case_ , unittest.TestCase ): """simple docstring""" a_ =GPTaTokenizer a_ =GPTaTokenizerFast a_ =True a_ ={'''add_prefix_space''': True} a_ =False def _lowercase ( self : List[str] ) -> List[Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowerCamelCase : Optional[int] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] __lowerCamelCase : List[Any] = dict(zip(__snake_case , range(len(__snake_case ) ) ) ) __lowerCamelCase : List[Any] = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __lowerCamelCase : Any = {'unk_token': '<unk>'} __lowerCamelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __lowerCamelCase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__snake_case ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__snake_case ) ) def _lowercase ( self : Union[str, Any] , **_a : Dict ) -> Any: kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **__snake_case ) def _lowercase ( self : List[Any] , **_a : Optional[Any] ) -> Optional[Any]: kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **__snake_case ) def _lowercase ( self : Tuple , _a : str ) -> Dict: __lowerCamelCase : str = 'lower newer' __lowerCamelCase : int = 'lower newer' return input_text, output_text def _lowercase ( self : List[Any] ) -> Any: __lowerCamelCase : Any = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __lowerCamelCase : Tuple = 'lower newer' __lowerCamelCase : Optional[Any] = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] __lowerCamelCase : str = tokenizer.tokenize(__snake_case , add_prefix_space=__snake_case ) self.assertListEqual(__snake_case , __snake_case ) __lowerCamelCase : Any = tokens + [tokenizer.unk_token] __lowerCamelCase : Tuple = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(__snake_case ) , __snake_case ) def _lowercase ( self : Dict ) -> List[Any]: if not self.test_rust_tokenizer: return __lowerCamelCase : int = self.get_tokenizer() __lowerCamelCase : int = self.get_rust_tokenizer(add_prefix_space=__snake_case ) __lowerCamelCase : Optional[Any] = 'lower newer' # Testing tokenization __lowerCamelCase : Dict = tokenizer.tokenize(__snake_case , add_prefix_space=__snake_case ) __lowerCamelCase : str = rust_tokenizer.tokenize(__snake_case ) self.assertListEqual(__snake_case , __snake_case ) # Testing conversion to ids without special tokens __lowerCamelCase : List[str] = tokenizer.encode(__snake_case , add_special_tokens=__snake_case , add_prefix_space=__snake_case ) __lowerCamelCase : Any = rust_tokenizer.encode(__snake_case , add_special_tokens=__snake_case ) self.assertListEqual(__snake_case , __snake_case ) # Testing conversion to ids with special tokens __lowerCamelCase : Optional[Any] = self.get_rust_tokenizer(add_prefix_space=__snake_case ) __lowerCamelCase : List[Any] = tokenizer.encode(__snake_case , add_prefix_space=__snake_case ) __lowerCamelCase : int = rust_tokenizer.encode(__snake_case ) self.assertListEqual(__snake_case , __snake_case ) # Testing the unknown token __lowerCamelCase : Union[str, Any] = tokens + [rust_tokenizer.unk_token] __lowerCamelCase : List[str] = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(__snake_case ) , __snake_case ) def _lowercase ( self : Union[str, Any] , *_a : int , **_a : List[str] ) -> Dict: # It's very difficult to mix/test pretokenization with byte-level # And get both GPT2 and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def _lowercase ( self : List[Any] , _a : List[str]=15 ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): __lowerCamelCase : List[Any] = self.rust_tokenizer_class.from_pretrained(__snake_case , **__snake_case ) # Simple input __lowerCamelCase : Union[str, Any] = 'This is a simple input' __lowerCamelCase : List[Any] = ['This is a simple input 1', 'This is a simple input 2'] __lowerCamelCase : Optional[int] = ('This is a simple input', 'This is a pair') __lowerCamelCase : str = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(__snake_case , tokenizer_r.encode , __snake_case , max_length=__snake_case , padding='max_length' ) # Simple input self.assertRaises(__snake_case , tokenizer_r.encode_plus , __snake_case , max_length=__snake_case , padding='max_length' ) # Simple input self.assertRaises( __snake_case , tokenizer_r.batch_encode_plus , __snake_case , max_length=__snake_case , padding='max_length' , ) # Pair input self.assertRaises(__snake_case , tokenizer_r.encode , __snake_case , max_length=__snake_case , padding='max_length' ) # Pair input self.assertRaises(__snake_case , tokenizer_r.encode_plus , __snake_case , max_length=__snake_case , padding='max_length' ) # Pair input self.assertRaises( __snake_case , tokenizer_r.batch_encode_plus , __snake_case , max_length=__snake_case , padding='max_length' , ) def _lowercase ( self : Union[str, Any] ) -> Tuple: __lowerCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>' ) # Simple input __lowerCamelCase : Any = 'This is a simple input' __lowerCamelCase : str = ['This is a simple input looooooooong', 'This is a simple input'] __lowerCamelCase : str = ('This is a simple input', 'This is a pair') __lowerCamelCase : List[str] = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] __lowerCamelCase : Dict = tokenizer.pad_token_id __lowerCamelCase : int = tokenizer(__snake_case , padding='max_length' , max_length=30 , return_tensors='np' ) __lowerCamelCase : int = tokenizer(__snake_case , padding=__snake_case , truncate=__snake_case , return_tensors='np' ) __lowerCamelCase : Dict = tokenizer(*__snake_case , padding='max_length' , max_length=60 , return_tensors='np' ) __lowerCamelCase : Any = tokenizer(__snake_case , padding=__snake_case , truncate=__snake_case , return_tensors='np' ) # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['input_ids'] ) self.assertTrue(0 in out_s['attention_mask'] ) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0] ) self.assertFalse(0 in out_sa['attention_mask'][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1] ) self.assertTrue(0 in out_sa['attention_mask'][1] ) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['input_ids'] ) self.assertTrue(0 in out_p['attention_mask'] ) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0] ) self.assertFalse(0 in out_pa['attention_mask'][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1] ) self.assertTrue(0 in out_pa['attention_mask'][1] ) def _lowercase ( self : str ) -> Any: __lowerCamelCase : Optional[Any] = '$$$' __lowerCamelCase : Dict = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=__snake_case , add_bos_token=__snake_case ) __lowerCamelCase : List[Any] = 'This is a simple input' __lowerCamelCase : Union[str, Any] = ['This is a simple input 1', 'This is a simple input 2'] __lowerCamelCase : List[Any] = tokenizer.bos_token_id __lowerCamelCase : Optional[Any] = tokenizer(__snake_case ) __lowerCamelCase : Any = tokenizer(__snake_case ) self.assertEqual(out_s.input_ids[0] , __snake_case ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) __lowerCamelCase : Tuple = tokenizer.decode(out_s.input_ids ) __lowerCamelCase : Optional[int] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , __snake_case ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def _lowercase ( self : str ) -> int: pass def _lowercase ( self : Dict ) -> List[str]: # TODO: change to self.get_tokenizers() when the fast version is implemented __lowerCamelCase : Any = [self.get_tokenizer(do_lower_case=__snake_case , add_bos_token=__snake_case )] for tokenizer in tokenizers: with self.subTest(f'{tokenizer.__class__.__name__}' ): __lowerCamelCase : Optional[int] = 'Encode this.' __lowerCamelCase : str = 'This one too please.' __lowerCamelCase : Any = tokenizer.encode(__snake_case , add_special_tokens=__snake_case ) encoded_sequence += tokenizer.encode(__snake_case , add_special_tokens=__snake_case ) __lowerCamelCase : Any = tokenizer.encode_plus( __snake_case , __snake_case , add_special_tokens=__snake_case , return_special_tokens_mask=__snake_case , ) __lowerCamelCase : Tuple = encoded_sequence_dict['input_ids'] __lowerCamelCase : Optional[int] = encoded_sequence_dict['special_tokens_mask'] self.assertEqual(len(__snake_case ) , len(__snake_case ) ) __lowerCamelCase : Dict = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(__snake_case ) ] __lowerCamelCase : Tuple = [x for x in filtered_sequence if x is not None] self.assertEqual(__snake_case , __snake_case ) @require_tokenizers class lowerCamelCase_ ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : List[str] ) -> Tuple: # More context: # https://huggingface.co/wjmcat/opt-350m-paddle/discussions/1 # https://huggingface.slack.com/archives/C01N44FJDHT/p1653511495183519 # https://github.com/huggingface/transformers/pull/17088#discussion_r871246439 __lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=__snake_case ) __lowerCamelCase : Union[str, Any] = 'A photo of a cat' __lowerCamelCase : List[str] = tokenizer.encode( __snake_case , ) self.assertEqual(__snake_case , [2, 250, 1345, 9, 10, 4758] ) tokenizer.save_pretrained('test_opt' ) __lowerCamelCase : str = AutoTokenizer.from_pretrained('./test_opt' ) __lowerCamelCase : int = tokenizer.encode( __snake_case , ) self.assertEqual(__snake_case , [2, 250, 1345, 9, 10, 4758] ) def _lowercase ( self : Any ) -> str: __lowerCamelCase : Any = AutoTokenizer.from_pretrained('facebook/opt-350m' , use_slow=__snake_case ) __lowerCamelCase : Optional[int] = 'A photo of a cat' __lowerCamelCase : Optional[Any] = tokenizer.encode( __snake_case , ) # Same as above self.assertEqual(__snake_case , [2, 250, 1345, 9, 10, 4758] ) @unittest.skip('This test is failing because of a bug in the fast tokenizer' ) def _lowercase ( self : Union[str, Any] ) -> str: __lowerCamelCase : int = AutoTokenizer.from_pretrained('facebook/opt-350m' , from_slow=__snake_case ) __lowerCamelCase : Optional[int] = 'bos' __lowerCamelCase : List[Any] = tokenizer.get_vocab()['bos'] __lowerCamelCase : List[str] = 'A photo of a cat' __lowerCamelCase : List[Any] = tokenizer.encode( __snake_case , ) # We changed the bos token self.assertEqual(__snake_case , [3_1957, 250, 1345, 9, 10, 4758] ) tokenizer.save_pretrained('./tok' ) __lowerCamelCase : Tuple = AutoTokenizer.from_pretrained('./tok' ) self.assertTrue(tokenizer.is_fast ) __lowerCamelCase : List[str] = tokenizer.encode( __snake_case , ) self.assertEqual(__snake_case , [3_1957, 250, 1345, 9, 10, 4758] )
208
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A__ : List[Any] =logging.get_logger(__name__) A__ : Any =torch.device('''cpu''') def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" _lowerCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = dct.pop(lowerCAmelCase ) _lowerCAmelCase = val def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [] for k in state_dict.keys(): _lowerCAmelCase = k if ".pwconv" in k: _lowerCAmelCase = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _lowerCAmelCase = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _lowerCAmelCase = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _lowerCAmelCase = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _lowerCAmelCase = k_new.split(""".""" ) if ls[2].isdigit(): _lowerCAmelCase = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _lowerCAmelCase = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _lowerCAmelCase = 10_00 _lowerCAmelCase = """huggingface/label-files""" _lowerCAmelCase = """imagenet-1k-id2label.json""" _lowerCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) _lowerCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} _lowerCAmelCase = idalabel _lowerCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _lowerCAmelCase = [3, 3, 6, 4] _lowerCAmelCase = [48, 56, 1_12, 2_20] elif swiftformer_name == "swiftformer_s": _lowerCAmelCase = [3, 3, 9, 6] _lowerCAmelCase = [48, 64, 1_68, 2_24] elif swiftformer_name == "swiftformer_l1": _lowerCAmelCase = [4, 3, 10, 5] _lowerCAmelCase = [48, 96, 1_92, 3_84] elif swiftformer_name == "swiftformer_l3": _lowerCAmelCase = [4, 4, 12, 6] _lowerCAmelCase = [64, 1_28, 3_20, 5_12] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _lowerCAmelCase = torch.hub.load_state_dict_from_url(lowerCAmelCase , map_location="""cpu""" , check_hash=lowerCAmelCase ) else: _lowerCAmelCase = torch.load(lowerCAmelCase , map_location="""cpu""" ) _lowerCAmelCase = checkpoint _lowerCAmelCase = create_rename_keys(lowerCAmelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model _lowerCAmelCase = SwiftFormerForImageClassification(lowerCAmelCase ).eval() hf_model.load_state_dict(lowerCAmelCase ) # prepare test inputs _lowerCAmelCase = prepare_img() _lowerCAmelCase = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _lowerCAmelCase = processor(images=lowerCAmelCase , return_tensors="""pt""" ) # compare outputs from both models _lowerCAmelCase = get_expected_output(lowerCAmelCase ) _lowerCAmelCase = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 10_00] ) assert torch.allclose(hf_logits[0, 0:5] , lowerCAmelCase , atol=1e-3 ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(f"Saving model {swiftformer_name} to {pytorch_dump_folder_path}" ) hf_model.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": A__ : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swiftformer_name''', default='''swiftformer_xs''', choices=['''swiftformer_xs''', '''swiftformer_s''', '''swiftformer_l1''', '''swiftformer_l3'''], type=str, help='''Name of the SwiftFormer model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''./converted_outputs/''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--original_ckpt''', default=None, type=str, help='''Path to the original model checkpoint.''') A__ : Tuple =parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
70
0
'''simple docstring''' import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging __snake_case = logging.get_logger(__name__) __snake_case = { '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/config.json''', # See all BART models at https://huggingface.co/models?filter=bart } class lowercase ( snake_case_ ): """simple docstring""" _a = '''bart''' _a = ['''past_key_values'''] _a = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self , UpperCamelCase_=50265 , UpperCamelCase_=1024 , UpperCamelCase_=12 , UpperCamelCase_=4096 , UpperCamelCase_=16 , UpperCamelCase_=12 , UpperCamelCase_=4096 , UpperCamelCase_=16 , UpperCamelCase_=0.0 , UpperCamelCase_=0.0 , UpperCamelCase_="gelu" , UpperCamelCase_=1024 , UpperCamelCase_=0.1 , UpperCamelCase_=0.0 , UpperCamelCase_=0.0 , UpperCamelCase_=0.02 , UpperCamelCase_=0.0 , UpperCamelCase_=False , UpperCamelCase_=True , UpperCamelCase_=3 , UpperCamelCase_=1 , UpperCamelCase_=0 , UpperCamelCase_=2 , UpperCamelCase_=True , UpperCamelCase_=2 , UpperCamelCase_=2 , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :Optional[int] = vocab_size UpperCamelCase__ :Union[str, Any] = max_position_embeddings UpperCamelCase__ :List[Any] = d_model UpperCamelCase__ :Union[str, Any] = encoder_ffn_dim UpperCamelCase__ :str = encoder_layers UpperCamelCase__ :Dict = encoder_attention_heads UpperCamelCase__ :Union[str, Any] = decoder_ffn_dim UpperCamelCase__ :Any = decoder_layers UpperCamelCase__ :List[Any] = decoder_attention_heads UpperCamelCase__ :List[str] = dropout UpperCamelCase__ :List[Any] = attention_dropout UpperCamelCase__ :Any = activation_dropout UpperCamelCase__ :List[Any] = activation_function UpperCamelCase__ :Tuple = init_std UpperCamelCase__ :List[Any] = encoder_layerdrop UpperCamelCase__ :Tuple = decoder_layerdrop UpperCamelCase__ :Any = classifier_dropout UpperCamelCase__ :Dict = use_cache UpperCamelCase__ :Any = encoder_layers UpperCamelCase__ :List[Any] = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=__snake_case , pad_token_id=__snake_case , bos_token_id=__snake_case , eos_token_id=__snake_case , is_encoder_decoder=__snake_case , decoder_start_token_id=__snake_case , forced_eos_token_id=__snake_case , **__snake_case , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , __snake_case ): UpperCamelCase__ :int = self.bos_token_id warnings.warn( F'''Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ''' '''The config can simply be saved and uploaded again to be fixed.''' ) class lowercase ( snake_case_ ): """simple docstring""" @property def lowerCAmelCase__ ( self ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: UpperCamelCase__ :List[str] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: UpperCamelCase__ :int = {0: '''batch'''} UpperCamelCase__ :Optional[int] = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: UpperCamelCase__ :Optional[Any] = {0: '''batch''', 1: '''decoder_sequence'''} UpperCamelCase__ :str = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(__snake_case , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. UpperCamelCase__ :Optional[Any] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: UpperCamelCase__ , UpperCamelCase__ :Union[str, Any] = self.num_layers for i in range(__snake_case ): UpperCamelCase__ :Tuple = {0: '''batch''', 2: '''past_sequence + sequence'''} UpperCamelCase__ :Dict = {0: '''batch''', 2: '''past_sequence + sequence'''} else: UpperCamelCase__ :int = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property def lowerCAmelCase__ ( self ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: UpperCamelCase__ :str = super().outputs else: UpperCamelCase__ :int = super(__snake_case , self ).outputs if self.use_past: UpperCamelCase__ , UpperCamelCase__ :Tuple = self.num_layers for i in range(__snake_case ): UpperCamelCase__ :int = {0: '''batch''', 2: '''past_sequence + sequence'''} UpperCamelCase__ :Optional[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = -1 , UpperCamelCase_ = -1 , UpperCamelCase_ = False , UpperCamelCase_ = None , ): '''simple docstring''' UpperCamelCase__ :List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) # Generate decoder inputs UpperCamelCase__ :Tuple = seq_length if not self.use_past else 1 UpperCamelCase__ :List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) UpperCamelCase__ :str = {F'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} UpperCamelCase__ :Optional[Any] = dict(**__snake_case , **__snake_case ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch UpperCamelCase__ , UpperCamelCase__ :Tuple = common_inputs['''input_ids'''].shape UpperCamelCase__ :Union[str, Any] = common_inputs['''decoder_input_ids'''].shape[1] UpperCamelCase__ , UpperCamelCase__ :Union[str, Any] = self.num_attention_heads UpperCamelCase__ :List[str] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCamelCase__ :Union[str, Any] = decoder_seq_length + 3 UpperCamelCase__ :Tuple = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) UpperCamelCase__ :str = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(__snake_case , __snake_case )] , dim=1 ) UpperCamelCase__ :Tuple = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered UpperCamelCase__ , UpperCamelCase__ :Optional[Any] = self.num_layers UpperCamelCase__ :Optional[Any] = min(__snake_case , __snake_case ) UpperCamelCase__ :Optional[Any] = max(__snake_case , __snake_case ) - min_num_layers UpperCamelCase__ :List[Any] = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(__snake_case ): common_inputs["past_key_values"].append( ( torch.zeros(__snake_case ), torch.zeros(__snake_case ), torch.zeros(__snake_case ), torch.zeros(__snake_case ), ) ) # TODO: test this. UpperCamelCase__ :Optional[Any] = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(__snake_case , __snake_case ): common_inputs["past_key_values"].append((torch.zeros(__snake_case ), torch.zeros(__snake_case )) ) return common_inputs def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = -1 , UpperCamelCase_ = -1 , UpperCamelCase_ = False , UpperCamelCase_ = None , ): '''simple docstring''' UpperCamelCase__ :List[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch UpperCamelCase__ , UpperCamelCase__ :Any = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values UpperCamelCase__ :Optional[int] = seqlen + 2 UpperCamelCase__ , UpperCamelCase__ :Tuple = self.num_layers UpperCamelCase__ , UpperCamelCase__ :int = self.num_attention_heads UpperCamelCase__ :str = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCamelCase__ :Union[str, Any] = common_inputs['''attention_mask'''].dtype UpperCamelCase__ :Tuple = torch.cat( [common_inputs['''attention_mask'''], torch.ones(__snake_case , __snake_case , dtype=__snake_case )] , dim=1 ) UpperCamelCase__ :Tuple = [ (torch.zeros(__snake_case ), torch.zeros(__snake_case )) for _ in range(__snake_case ) ] return common_inputs def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = -1 , UpperCamelCase_ = -1 , UpperCamelCase_ = False , UpperCamelCase_ = None , ): '''simple docstring''' UpperCamelCase__ :List[str] = compute_effective_axis_dimension( __snake_case , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCamelCase__ :List[str] = tokenizer.num_special_tokens_to_add(__snake_case ) UpperCamelCase__ :List[Any] = compute_effective_axis_dimension( __snake_case , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__snake_case ) # Generate dummy inputs according to compute batch and sequence UpperCamelCase__ :List[str] = [''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size UpperCamelCase__ :str = dict(tokenizer(__snake_case , return_tensors=__snake_case ) ) return common_inputs def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = -1 , UpperCamelCase_ = -1 , UpperCamelCase_ = False , UpperCamelCase_ = None , ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: UpperCamelCase__ :int = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __snake_case , batch_size=__snake_case , seq_length=__snake_case , is_pair=__snake_case , framework=__snake_case ) elif self.task == "causal-lm": UpperCamelCase__ :Tuple = self._generate_dummy_inputs_for_causal_lm( __snake_case , batch_size=__snake_case , seq_length=__snake_case , is_pair=__snake_case , framework=__snake_case ) else: UpperCamelCase__ :Optional[int] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , batch_size=__snake_case , seq_length=__snake_case , is_pair=__snake_case , framework=__snake_case ) return common_inputs def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: UpperCamelCase__ :int = super()._flatten_past_key_values_(__snake_case , __snake_case , __snake_case , __snake_case ) else: UpperCamelCase__ :Optional[Any] = super(__snake_case , self )._flatten_past_key_values_( __snake_case , __snake_case , __snake_case , __snake_case )
97
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) A__ : List[Any] =pytest.mark.integration @pytest.mark.parametrize("""path""" , ["""paws""", """csv"""] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" inspect_dataset(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = path + """.py""" assert script_name in os.listdir(lowerCAmelCase ) assert "__pycache__" not in os.listdir(lowerCAmelCase ) @pytest.mark.filterwarnings("""ignore:inspect_metric is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.parametrize("""path""" , ["""accuracy"""] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" inspect_metric(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = path + """.py""" assert script_name in os.listdir(lowerCAmelCase ) assert "__pycache__" not in os.listdir(lowerCAmelCase ) @pytest.mark.parametrize( """path, config_name, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_config_info(lowerCAmelCase , config_name=lowerCAmelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with pytest.raises(lowerCAmelCase ): get_dataset_config_info(lowerCAmelCase , config_name=lowerCAmelCase ) @pytest.mark.parametrize( """path, expected""" , [ ("""squad""", """plain_text"""), ("""acronym_identification""", """default"""), ("""lhoestq/squad""", """plain_text"""), ("""lhoestq/test""", """default"""), ("""lhoestq/demo1""", """lhoestq--demo1"""), ("""dalle-mini/wit""", """dalle-mini--wit"""), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_config_names(lowerCAmelCase ) assert expected in config_names @pytest.mark.parametrize( """path, expected_configs, expected_splits_in_first_config""" , [ ("""squad""", ["""plain_text"""], ["""train""", """validation"""]), ("""dalle-mini/wit""", ["""dalle-mini--wit"""], ["""train"""]), ("""paws""", ["""labeled_final""", """labeled_swap""", """unlabeled_final"""], ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_infos(lowerCAmelCase ) assert list(infos.keys() ) == expected_configs _lowerCAmelCase = expected_configs[0] assert expected_config in infos _lowerCAmelCase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( """path, expected_config, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_infos(lowerCAmelCase ) assert expected_config in infos _lowerCAmelCase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with pytest.raises(lowerCAmelCase ): get_dataset_split_names(lowerCAmelCase , config_name=lowerCAmelCase )
70
0
"""simple docstring""" def _A ( lowercase ): """simple docstring""" a =generate_pascal_triangle(lowercase ) for row_idx in range(lowercase ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=''' ''' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=''' ''' ) else: print(triangle[row_idx][col_idx] , end='''''' ) print() def _A ( lowercase ): """simple docstring""" if not isinstance(lowercase , lowercase ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) a =[] for current_row_idx in range(lowercase ): a =populate_current_row(lowercase , lowercase ) triangle.append(lowercase ) return triangle def _A ( lowercase , lowercase ): """simple docstring""" a =[-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 a , a =1, 1 for current_col_idx in range(1 , lowercase ): calculate_current_element( lowercase , lowercase , lowercase , lowercase ) return current_row def _A ( lowercase , lowercase , lowercase , lowercase , ): """simple docstring""" a =triangle[current_row_idx - 1][current_col_idx - 1] a =triangle[current_row_idx - 1][current_col_idx] a =above_to_left_elt + above_to_right_elt def _A ( lowercase ): """simple docstring""" if not isinstance(lowercase , lowercase ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) a =[[1]] for row_index in range(1 , lowercase ): a =[0] + result[-1] + [0] a =row_index + 1 # Calculate the number of distinct elements in a row a =sum(divmod(lowercase , 2 ) ) a =[ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] a =row_first_half[: (row_index + 1) // 2] row_second_half.reverse() a =row_first_half + row_second_half result.append(lowercase ) return result def _A ( ): """simple docstring""" from collections.abc import Callable from timeit import timeit def benchmark_a_function(lowercase , lowercase ) -> None: a =f'''{func.__name__}({value})''' a =timeit(f'''__main__.{call}''' , setup='''import __main__''' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(f'''{call:38} -- {timing:.4f} seconds''' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(lowercase , lowercase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
81
'''simple docstring''' from torch import nn def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f"Unsupported activation function: {act_fn}" )
70
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() snake_case_ = logging.get_logger(__name__) def _lowerCAmelCase ( lowercase_ , lowercase_=False ): UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'deit.embeddings.cls_token'), ('dist_token', 'deit.embeddings.distillation_token'), ('patch_embed.proj.weight', 'deit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'deit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'deit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('deit' ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ('norm.weight', 'deit.layernorm.weight'), ('norm.bias', 'deit.layernorm.bias'), ('head.weight', 'cls_classifier.weight'), ('head.bias', 'cls_classifier.bias'), ('head_dist.weight', 'distillation_classifier.weight'), ('head_dist.bias', 'distillation_classifier.bias'), ] ) return rename_keys def _lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase = '' else: UpperCAmelCase = 'deit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_ ): UpperCAmelCase = dct.pop(lowercase_ ) UpperCAmelCase = val def _lowerCAmelCase ( ): UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' UpperCAmelCase = Image.open(requests.get(lowercase_ , stream=lowercase_ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( lowercase_ , lowercase_ ): UpperCAmelCase = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase = 1000 UpperCAmelCase = 'huggingface/label-files' UpperCAmelCase = 'imagenet-1k-id2label.json' UpperCAmelCase = json.load(open(hf_hub_download(lowercase_ , lowercase_ , repo_type='dataset' ) , 'r' ) ) UpperCAmelCase = {int(lowercase_ ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = int(deit_name[-6:-4] ) UpperCAmelCase = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith('tiny' ): UpperCAmelCase = 192 UpperCAmelCase = 768 UpperCAmelCase = 12 UpperCAmelCase = 3 elif deit_name[9:].startswith('small' ): UpperCAmelCase = 384 UpperCAmelCase = 1536 UpperCAmelCase = 12 UpperCAmelCase = 6 if deit_name[9:].startswith('base' ): pass elif deit_name[4:].startswith('large' ): UpperCAmelCase = 1024 UpperCAmelCase = 4096 UpperCAmelCase = 24 UpperCAmelCase = 16 # load original model from timm UpperCAmelCase = timm.create_model(lowercase_ , pretrained=lowercase_ ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase = timm_model.state_dict() UpperCAmelCase = create_rename_keys(lowercase_ , lowercase_ ) for src, dest in rename_keys: rename_key(lowercase_ , lowercase_ , lowercase_ ) read_in_q_k_v(lowercase_ , lowercase_ , lowercase_ ) # load HuggingFace model UpperCAmelCase = DeiTForImageClassificationWithTeacher(lowercase_ ).eval() model.load_state_dict(lowercase_ ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase = DeiTImageProcessor(size=lowercase_ , crop_size=config.image_size ) UpperCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ) UpperCAmelCase = encoding['pixel_values'] UpperCAmelCase = model(lowercase_ ) UpperCAmelCase = timm_model(lowercase_ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase_ , outputs.logits , atol=1e-3 ) Path(lowercase_ ).mkdir(exist_ok=lowercase_ ) print(F"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase_ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(lowercase_ ) if __name__ == "__main__": snake_case_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--deit_name""", default="""vit_deit_base_distilled_patch16_224""", type=str, help="""Name of the DeiT timm model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) snake_case_ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
78
'''simple docstring''' # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position A__ : Dict ='''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip A__ : Tuple =concatenate_datasets A__ : Dict =DownloadConfig A__ : int =DownloadManager A__ : Union[str, Any] =DownloadMode A__ : Tuple =DownloadConfig A__ : Optional[Any] =DownloadMode A__ : str =DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
70
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __SCREAMING_SNAKE_CASE ={ '''configuration_longt5''': ['''LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongT5Config''', '''LongT5OnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE =[ '''LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongT5EncoderModel''', '''LongT5ForConditionalGeneration''', '''LongT5Model''', '''LongT5PreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE =[ '''FlaxLongT5ForConditionalGeneration''', '''FlaxLongT5Model''', '''FlaxLongT5PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
213
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A__ : Tuple ={ '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : int =['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Any =[ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys A__ : Any =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
70
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE_ = { '''configuration_blip''': [ '''BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BlipConfig''', '''BlipTextConfig''', '''BlipVisionConfig''', ], '''processing_blip''': ['''BlipProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = ['''BlipImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ '''BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BlipModel''', '''BlipPreTrainedModel''', '''BlipForConditionalGeneration''', '''BlipForQuestionAnswering''', '''BlipVisionModel''', '''BlipTextModel''', '''BlipForImageTextRetrieval''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ '''TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFBlipModel''', '''TFBlipPreTrainedModel''', '''TFBlipForConditionalGeneration''', '''TFBlipForQuestionAnswering''', '''TFBlipVisionModel''', '''TFBlipTextModel''', '''TFBlipForImageTextRetrieval''', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys SCREAMING_SNAKE_CASE_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
296
'''simple docstring''' def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [0 for i in range(r + 1 )] # nc0 = 1 _lowerCAmelCase = 1 for i in range(1 , n + 1 ): # to compute current row from previous row. _lowerCAmelCase = min(lowerCAmelCase , lowerCAmelCase ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=10, r=5))
70
0
"""simple docstring""" from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch _A = logging.get_logger(__name__) class lowerCamelCase ( snake_case_ ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['''pixel_values'''] def __init__(self , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = PILImageResampling.BILINEAR , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = True , **_lowerCamelCase , ): """simple docstring""" super().__init__(**__snake_case ) UpperCAmelCase__ : Union[str, Any] = size if size is not None else {"""shortest_edge""": 224} UpperCAmelCase__ : List[str] = get_size_dict(__snake_case , default_to_square=__snake_case ) UpperCAmelCase__ : List[Any] = crop_size if crop_size is not None else {"""height""": 256, """width""": 256} UpperCAmelCase__ : Any = get_size_dict(__snake_case , param_name="""crop_size""" ) UpperCAmelCase__ : List[Any] = do_resize UpperCAmelCase__ : Optional[int] = size UpperCAmelCase__ : List[str] = resample UpperCAmelCase__ : str = do_rescale UpperCAmelCase__ : Union[str, Any] = rescale_factor UpperCAmelCase__ : Optional[int] = do_center_crop UpperCAmelCase__ : Optional[Any] = crop_size UpperCAmelCase__ : Union[str, Any] = do_flip_channel_order def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = PIL.Image.BILINEAR , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : List[str] = get_size_dict(__snake_case , default_to_square=__snake_case ) if "shortest_edge" not in size: raise ValueError(F"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCAmelCase__ : Tuple = get_resize_output_image_size(__snake_case , size=size["""shortest_edge"""] , default_to_square=__snake_case ) return resize(__snake_case , size=__snake_case , resample=__snake_case , data_format=__snake_case , **__snake_case ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Optional[int] = get_size_dict(__snake_case ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(__snake_case , size=(size["""height"""], size["""width"""]) , data_format=__snake_case , **__snake_case ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" return rescale(__snake_case , scale=__snake_case , data_format=__snake_case , **__snake_case ) def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" return flip_channel_order(__snake_case , data_format=__snake_case ) def _a (self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = ChannelDimension.FIRST , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : str = do_resize if do_resize is not None else self.do_resize UpperCAmelCase__ : List[str] = resample if resample is not None else self.resample UpperCAmelCase__ : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase__ : Optional[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase__ : int = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase__ : Union[str, Any] = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) UpperCAmelCase__ : Optional[int] = size if size is not None else self.size UpperCAmelCase__ : Optional[int] = get_size_dict(__snake_case , default_to_square=__snake_case ) UpperCAmelCase__ : Optional[int] = crop_size if crop_size is not None else self.crop_size UpperCAmelCase__ : Union[str, Any] = get_size_dict(__snake_case , param_name="""crop_size""" ) UpperCAmelCase__ : Optional[Any] = make_list_of_images(__snake_case ) if not valid_images(__snake_case ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. UpperCAmelCase__ : str = [to_numpy_array(__snake_case ) for image in images] if do_resize: UpperCAmelCase__ : List[Any] = [self.resize(image=__snake_case , size=__snake_case , resample=__snake_case ) for image in images] if do_center_crop: UpperCAmelCase__ : Optional[int] = [self.center_crop(image=__snake_case , size=__snake_case ) for image in images] if do_rescale: UpperCAmelCase__ : Any = [self.rescale(image=__snake_case , scale=__snake_case ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: UpperCAmelCase__ : str = [self.flip_channel_order(image=__snake_case ) for image in images] UpperCAmelCase__ : Union[str, Any] = [to_channel_dimension_format(__snake_case , __snake_case ) for image in images] UpperCAmelCase__ : int = {"""pixel_values""": images} return BatchFeature(data=__snake_case , tensor_type=__snake_case ) def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" UpperCAmelCase__ : int = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(__snake_case ) != len(__snake_case ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(__snake_case ): UpperCAmelCase__ : int = target_sizes.numpy() UpperCAmelCase__ : Union[str, Any] = [] for idx in range(len(__snake_case ) ): UpperCAmelCase__ : Optional[int] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=__snake_case ) UpperCAmelCase__ : Tuple = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(__snake_case ) else: UpperCAmelCase__ : str = logits.argmax(dim=1 ) UpperCAmelCase__ : List[str] = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
171
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Optional import pandas as pd import pyarrow as pa import datasets from datasets.table import table_cast @dataclass class UpperCAmelCase ( datasets.BuilderConfig ): _lowercase: Optional[datasets.Features] = None class UpperCAmelCase ( datasets.ArrowBasedBuilder ): _lowercase: Tuple = PandasConfig def lowercase__ ( self : Optional[Any] ) -> str: return datasets.DatasetInfo(features=self.config.features ) def lowercase__ ( self : List[str] , __snake_case : Dict ) -> int: if not self.config.data_files: raise ValueError(f"At least one data file must be specified, but got data_files={self.config.data_files}" ) _lowerCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(__snake_case , (str, list, tuple) ): _lowerCAmelCase = data_files if isinstance(__snake_case , __snake_case ): _lowerCAmelCase = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _lowerCAmelCase = [dl_manager.iter_files(__snake_case ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] _lowerCAmelCase = [] for split_name, files in data_files.items(): if isinstance(__snake_case , __snake_case ): _lowerCAmelCase = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive _lowerCAmelCase = [dl_manager.iter_files(__snake_case ) for file in files] splits.append(datasets.SplitGenerator(name=__snake_case , gen_kwargs={"""files""": files} ) ) return splits def lowercase__ ( self : List[Any] , __snake_case : pa.Table ) -> pa.Table: if self.config.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example _lowerCAmelCase = table_cast(__snake_case , self.config.features.arrow_schema ) return pa_table def lowercase__ ( self : Dict , __snake_case : Optional[Any] ) -> Any: for i, file in enumerate(itertools.chain.from_iterable(__snake_case ) ): with open(__snake_case , """rb""" ) as f: _lowerCAmelCase = pa.Table.from_pandas(pd.read_pickle(__snake_case ) ) yield i, self._cast_table(__snake_case )
70
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase = { '''configuration_whisper''': ['''WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''WhisperConfig''', '''WhisperOnnxConfig'''], '''feature_extraction_whisper''': ['''WhisperFeatureExtractor'''], '''processing_whisper''': ['''WhisperProcessor'''], '''tokenization_whisper''': ['''WhisperTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ['''WhisperTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''WhisperForConditionalGeneration''', '''WhisperModel''', '''WhisperPreTrainedModel''', '''WhisperForAudioClassification''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWhisperForConditionalGeneration''', '''TFWhisperModel''', '''TFWhisperPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''FlaxWhisperForConditionalGeneration''', '''FlaxWhisperModel''', '''FlaxWhisperPreTrainedModel''', '''FlaxWhisperForAudioClassification''', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
275
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class UpperCAmelCase : def __init__( self : str , __snake_case : Any ) -> str: _lowerCAmelCase = str(id_ ) _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = [] _lowerCAmelCase = {} # {vertex:distance} def __lt__( self : List[str] , __snake_case : Union[str, Any] ) -> Any: return self.key < other.key def __repr__( self : Optional[Any] ) -> Optional[Any]: return self.id def lowercase__ ( self : Union[str, Any] , __snake_case : Tuple ) -> Optional[Any]: self.neighbors.append(__snake_case ) def lowercase__ ( self : Tuple , __snake_case : List[str] , __snake_case : Tuple ) -> Any: _lowerCAmelCase = weight def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , lowerCAmelCase ) graph[b - 1].add_edge(graph[a - 1] , lowerCAmelCase ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [] for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = graph[:] while q: _lowerCAmelCase = min(lowerCAmelCase ) q.remove(lowerCAmelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] for i in range(1 , len(lowerCAmelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = list(lowerCAmelCase ) hq.heapify(lowerCAmelCase ) while h: _lowerCAmelCase = hq.heappop(lowerCAmelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] hq.heapify(lowerCAmelCase ) for i in range(1 , len(lowerCAmelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def UpperCamelCase__ ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
70
0
'''simple docstring''' import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __lowercase ( __lowercase ) -> Tuple: '''simple docstring''' _A = {} _A = tokenizer(example["content"] , truncation=__lowercase )["input_ids"] _A = len(example["content"] ) / len(output["input_ids"] ) return output lowerCamelCase_ = HfArgumentParser(PretokenizationArguments) lowerCamelCase_ = parser.parse_args() if args.num_workers is None: lowerCamelCase_ = multiprocessing.cpu_count() lowerCamelCase_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) lowerCamelCase_ = time.time() lowerCamelCase_ = load_dataset(args.dataset_name, split='''train''') print(F"""Dataset loaded in {time.time()-t_start:.2f}s""") lowerCamelCase_ = time.time() lowerCamelCase_ = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(F"""Dataset tokenized in {time.time()-t_start:.2f}s""") lowerCamelCase_ = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F"""Data pushed to the hub in {time.time()-t_start:.2f}s""")
79
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase ( unittest.TestCase ): @slow def lowercase__ ( self : List[Any] ) -> str: _lowerCAmelCase = AutoModelForSeqaSeqLM.from_pretrained("""google/mt5-small""" , return_dict=__snake_case ).to(__snake_case ) _lowerCAmelCase = AutoTokenizer.from_pretrained("""google/mt5-small""" ) _lowerCAmelCase = tokenizer("""Hello there""" , return_tensors="""pt""" ).input_ids _lowerCAmelCase = tokenizer("""Hi I am""" , return_tensors="""pt""" ).input_ids _lowerCAmelCase = model(input_ids.to(__snake_case ) , labels=labels.to(__snake_case ) ).loss _lowerCAmelCase = -(labels.shape[-1] * loss.item()) _lowerCAmelCase = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
70
0
import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class A_ ( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase_ (self ) -> Optional[Any]: __UpperCAmelCase = ['''a''', '''b''', '''c'''] # Defaults to last layer if both are None __UpperCAmelCase , __UpperCAmelCase = get_aligned_output_features_output_indices(__snake_case , __snake_case , __snake_case ) self.assertEqual(__snake_case , ['''c'''] ) self.assertEqual(__snake_case , [2] ) # Out indices set to match out features __UpperCAmelCase , __UpperCAmelCase = get_aligned_output_features_output_indices(['''a''', '''c'''] , __snake_case , __snake_case ) self.assertEqual(__snake_case , ['''a''', '''c'''] ) self.assertEqual(__snake_case , [0, 2] ) # Out features set to match out indices __UpperCAmelCase , __UpperCAmelCase = get_aligned_output_features_output_indices(__snake_case , [0, 2] , __snake_case ) self.assertEqual(__snake_case , ['''a''', '''c'''] ) self.assertEqual(__snake_case , [0, 2] ) # Out features selected from negative indices __UpperCAmelCase , __UpperCAmelCase = get_aligned_output_features_output_indices(__snake_case , [-3, -1] , __snake_case ) self.assertEqual(__snake_case , ['''a''', '''c'''] ) self.assertEqual(__snake_case , [-3, -1] ) def lowerCAmelCase_ (self ) -> Union[str, Any]: # Stage names must be set with self.assertRaises(__snake_case ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 1) , __snake_case ) # Out features must be a list with self.assertRaises(__snake_case ): verify_out_features_out_indices(('''a''', '''b''') , (0, 1) , ['''a''', '''b'''] ) # Out features must be a subset of stage names with self.assertRaises(__snake_case ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 1) , ['''a'''] ) # Out indices must be a list or tuple with self.assertRaises(__snake_case ): verify_out_features_out_indices(__snake_case , 0 , ['''a''', '''b'''] ) # Out indices must be a subset of stage names with self.assertRaises(__snake_case ): verify_out_features_out_indices(__snake_case , (0, 1) , ['''a'''] ) # Out features and out indices must be the same length with self.assertRaises(__snake_case ): verify_out_features_out_indices(['''a''', '''b'''] , (0,) , ['''a''', '''b''', '''c'''] ) # Out features should match out indices with self.assertRaises(__snake_case ): verify_out_features_out_indices(['''a''', '''b'''] , (0, 2) , ['''a''', '''b''', '''c'''] ) # Out features and out indices should be in order with self.assertRaises(__snake_case ): verify_out_features_out_indices(['''b''', '''a'''] , (0, 1) , ['''a''', '''b'''] ) # Check passes with valid inputs verify_out_features_out_indices(['''a''', '''b''', '''d'''] , (0, 1, -1) , ['''a''', '''b''', '''c''', '''d'''] ) def lowerCAmelCase_ (self ) -> List[str]: __UpperCAmelCase = BackboneMixin() __UpperCAmelCase = ['''a''', '''b''', '''c'''] __UpperCAmelCase = ['''a''', '''c'''] __UpperCAmelCase = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ['''a''', '''c'''] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly __UpperCAmelCase = ['''a''', '''b'''] self.assertEqual(backbone.out_features , ['''a''', '''b'''] ) self.assertEqual(backbone.out_indices , [0, 1] ) __UpperCAmelCase = [-3, -1] self.assertEqual(backbone.out_features , ['''a''', '''c'''] ) self.assertEqual(backbone.out_indices , [-3, -1] )
333
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging A__ : Any =logging.get_logger(__name__) A__ : List[Any] ='''▁''' A__ : Optional[int] ={'''vocab_file''': '''sentencepiece.bpe.model'''} A__ : Union[str, Any] ={ '''vocab_file''': { '''facebook/nllb-200-distilled-600M''': ( '''https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model''' ), } } A__ : Dict ={ '''facebook/nllb-200-distilled-600M''': 10_24, } # fmt: off A__ : Union[str, Any] =['''ace_Arab''', '''ace_Latn''', '''acm_Arab''', '''acq_Arab''', '''aeb_Arab''', '''afr_Latn''', '''ajp_Arab''', '''aka_Latn''', '''amh_Ethi''', '''apc_Arab''', '''arb_Arab''', '''ars_Arab''', '''ary_Arab''', '''arz_Arab''', '''asm_Beng''', '''ast_Latn''', '''awa_Deva''', '''ayr_Latn''', '''azb_Arab''', '''azj_Latn''', '''bak_Cyrl''', '''bam_Latn''', '''ban_Latn''', '''bel_Cyrl''', '''bem_Latn''', '''ben_Beng''', '''bho_Deva''', '''bjn_Arab''', '''bjn_Latn''', '''bod_Tibt''', '''bos_Latn''', '''bug_Latn''', '''bul_Cyrl''', '''cat_Latn''', '''ceb_Latn''', '''ces_Latn''', '''cjk_Latn''', '''ckb_Arab''', '''crh_Latn''', '''cym_Latn''', '''dan_Latn''', '''deu_Latn''', '''dik_Latn''', '''dyu_Latn''', '''dzo_Tibt''', '''ell_Grek''', '''eng_Latn''', '''epo_Latn''', '''est_Latn''', '''eus_Latn''', '''ewe_Latn''', '''fao_Latn''', '''pes_Arab''', '''fij_Latn''', '''fin_Latn''', '''fon_Latn''', '''fra_Latn''', '''fur_Latn''', '''fuv_Latn''', '''gla_Latn''', '''gle_Latn''', '''glg_Latn''', '''grn_Latn''', '''guj_Gujr''', '''hat_Latn''', '''hau_Latn''', '''heb_Hebr''', '''hin_Deva''', '''hne_Deva''', '''hrv_Latn''', '''hun_Latn''', '''hye_Armn''', '''ibo_Latn''', '''ilo_Latn''', '''ind_Latn''', '''isl_Latn''', '''ita_Latn''', '''jav_Latn''', '''jpn_Jpan''', '''kab_Latn''', '''kac_Latn''', '''kam_Latn''', '''kan_Knda''', '''kas_Arab''', '''kas_Deva''', '''kat_Geor''', '''knc_Arab''', '''knc_Latn''', '''kaz_Cyrl''', '''kbp_Latn''', '''kea_Latn''', '''khm_Khmr''', '''kik_Latn''', '''kin_Latn''', '''kir_Cyrl''', '''kmb_Latn''', '''kon_Latn''', '''kor_Hang''', '''kmr_Latn''', '''lao_Laoo''', '''lvs_Latn''', '''lij_Latn''', '''lim_Latn''', '''lin_Latn''', '''lit_Latn''', '''lmo_Latn''', '''ltg_Latn''', '''ltz_Latn''', '''lua_Latn''', '''lug_Latn''', '''luo_Latn''', '''lus_Latn''', '''mag_Deva''', '''mai_Deva''', '''mal_Mlym''', '''mar_Deva''', '''min_Latn''', '''mkd_Cyrl''', '''plt_Latn''', '''mlt_Latn''', '''mni_Beng''', '''khk_Cyrl''', '''mos_Latn''', '''mri_Latn''', '''zsm_Latn''', '''mya_Mymr''', '''nld_Latn''', '''nno_Latn''', '''nob_Latn''', '''npi_Deva''', '''nso_Latn''', '''nus_Latn''', '''nya_Latn''', '''oci_Latn''', '''gaz_Latn''', '''ory_Orya''', '''pag_Latn''', '''pan_Guru''', '''pap_Latn''', '''pol_Latn''', '''por_Latn''', '''prs_Arab''', '''pbt_Arab''', '''quy_Latn''', '''ron_Latn''', '''run_Latn''', '''rus_Cyrl''', '''sag_Latn''', '''san_Deva''', '''sat_Beng''', '''scn_Latn''', '''shn_Mymr''', '''sin_Sinh''', '''slk_Latn''', '''slv_Latn''', '''smo_Latn''', '''sna_Latn''', '''snd_Arab''', '''som_Latn''', '''sot_Latn''', '''spa_Latn''', '''als_Latn''', '''srd_Latn''', '''srp_Cyrl''', '''ssw_Latn''', '''sun_Latn''', '''swe_Latn''', '''swh_Latn''', '''szl_Latn''', '''tam_Taml''', '''tat_Cyrl''', '''tel_Telu''', '''tgk_Cyrl''', '''tgl_Latn''', '''tha_Thai''', '''tir_Ethi''', '''taq_Latn''', '''taq_Tfng''', '''tpi_Latn''', '''tsn_Latn''', '''tso_Latn''', '''tuk_Latn''', '''tum_Latn''', '''tur_Latn''', '''twi_Latn''', '''tzm_Tfng''', '''uig_Arab''', '''ukr_Cyrl''', '''umb_Latn''', '''urd_Arab''', '''uzn_Latn''', '''vec_Latn''', '''vie_Latn''', '''war_Latn''', '''wol_Latn''', '''xho_Latn''', '''ydd_Hebr''', '''yor_Latn''', '''yue_Hant''', '''zho_Hans''', '''zho_Hant''', '''zul_Latn'''] class UpperCAmelCase ( snake_case_ ): _lowercase: int = VOCAB_FILES_NAMES _lowercase: str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase: Dict = PRETRAINED_VOCAB_FILES_MAP _lowercase: str = ['''input_ids''', '''attention_mask'''] _lowercase: List[int] = [] _lowercase: List[int] = [] def __init__( self : int , __snake_case : Optional[Any] , __snake_case : Dict="<s>" , __snake_case : Optional[int]="</s>" , __snake_case : Dict="</s>" , __snake_case : str="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : List[Any]=None , __snake_case : Union[str, Any]=None , __snake_case : int=None , __snake_case : Optional[Dict[str, Any]] = None , __snake_case : str=None , __snake_case : str=False , **__snake_case : List[Any] , ) -> List[str]: # Mask token behave like a normal word, i.e. include the space before it _lowerCAmelCase = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token _lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase = legacy_behaviour super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , tokenizer_file=__snake_case , src_lang=__snake_case , tgt_lang=__snake_case , additional_special_tokens=__snake_case , sp_model_kwargs=self.sp_model_kwargs , legacy_behaviour=__snake_case , **__snake_case , ) _lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) _lowerCAmelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token _lowerCAmelCase = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _lowerCAmelCase = 1 _lowerCAmelCase = len(self.sp_model ) _lowerCAmelCase = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(__snake_case ) } _lowerCAmelCase = {v: k for k, v in self.lang_code_to_id.items()} _lowerCAmelCase = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) _lowerCAmelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} _lowerCAmelCase = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) _lowerCAmelCase = src_lang if src_lang is not None else """eng_Latn""" _lowerCAmelCase = self.lang_code_to_id[self._src_lang] _lowerCAmelCase = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : List[str] ) -> List[str]: _lowerCAmelCase = self.__dict__.copy() _lowerCAmelCase = None _lowerCAmelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self : Dict , __snake_case : Optional[Any] ) -> Dict: _lowerCAmelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _lowerCAmelCase = {} _lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def lowercase__ ( self : List[Any] ) -> Any: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def lowercase__ ( self : int ) -> str: return self._src_lang @src_lang.setter def lowercase__ ( self : Dict , __snake_case : str ) -> None: _lowerCAmelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase__ ( self : List[Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) _lowerCAmelCase = [1] * len(self.prefix_tokens ) _lowerCAmelCase = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__snake_case )) + suffix_ones return prefix_ones + ([0] * len(__snake_case )) + ([0] * len(__snake_case )) + suffix_ones def lowercase__ ( self : Optional[Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase__ ( self : Optional[Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: _lowerCAmelCase = [self.sep_token_id] _lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase__ ( self : List[str] , __snake_case : Union[str, Any] , __snake_case : str , __snake_case : Optional[str] , __snake_case : Optional[str] , **__snake_case : Optional[int] ) -> Dict: if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) _lowerCAmelCase = src_lang _lowerCAmelCase = self(__snake_case , add_special_tokens=__snake_case , return_tensors=__snake_case , **__snake_case ) _lowerCAmelCase = self.convert_tokens_to_ids(__snake_case ) _lowerCAmelCase = tgt_lang_id return inputs def lowercase__ ( self : List[Any] ) -> Optional[int]: _lowerCAmelCase = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase__ ( self : Optional[int] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def lowercase__ ( self : Optional[Any] , __snake_case : Union[str, Any] ) -> Tuple: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _lowerCAmelCase = self.sp_model.PieceToId(__snake_case ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowercase__ ( self : List[Any] , __snake_case : Union[str, Any] ) -> Optional[int]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowercase__ ( self : Optional[int] , __snake_case : Union[str, Any] ) -> str: _lowerCAmelCase = """""".join(__snake_case ).replace(__snake_case , """ """ ).strip() return out_string def lowercase__ ( self : str , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return _lowerCAmelCase = os.path.join( __snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , """wb""" ) as fi: _lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,) def lowercase__ ( self : Optional[Any] , __snake_case : List[str] , __snake_case : str = "eng_Latn" , __snake_case : Optional[List[str]] = None , __snake_case : str = "fra_Latn" , **__snake_case : Optional[int] , ) -> BatchEncoding: _lowerCAmelCase = src_lang _lowerCAmelCase = tgt_lang return super().prepare_seqaseq_batch(__snake_case , __snake_case , **__snake_case ) def lowercase__ ( self : str ) -> Tuple: return self.set_src_lang_special_tokens(self.src_lang ) def lowercase__ ( self : Dict ) -> Optional[Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase__ ( self : str , __snake_case : int ) -> None: _lowerCAmelCase = self.lang_code_to_id[src_lang] if self.legacy_behaviour: _lowerCAmelCase = [] _lowerCAmelCase = [self.eos_token_id, self.cur_lang_code] else: _lowerCAmelCase = [self.cur_lang_code] _lowerCAmelCase = [self.eos_token_id] def lowercase__ ( self : Any , __snake_case : str ) -> None: _lowerCAmelCase = self.lang_code_to_id[lang] if self.legacy_behaviour: _lowerCAmelCase = [] _lowerCAmelCase = [self.eos_token_id, self.cur_lang_code] else: _lowerCAmelCase = [self.cur_lang_code] _lowerCAmelCase = [self.eos_token_id]
70
0
import json import os from pathlib import Path import pytest from datasets.download.download_config import DownloadConfig from datasets.download.download_manager import DownloadManager from datasets.utils.file_utils import hash_url_to_filename __A : Any = '''http://www.mocksite.com/file1.txt''' __A : List[str] = '''"text": ["foo", "foo"]''' __A : Dict = '''6d8ce9aa78a471c7477201efbeabd3bb01ac2e7d100a6dc024ba1608361f90a8''' class A_ : UpperCAmelCase__ = 2_0_0 UpperCAmelCase__ = {'''Content-Length''': '''100'''} UpperCAmelCase__ = {} def _lowercase ( self , **_A ): '''simple docstring''' return [bytes(__snake_case , '''utf-8''' )] def __SCREAMING_SNAKE_CASE ( *UpperCamelCase__ , **UpperCamelCase__ ) -> Tuple: '''simple docstring''' return MockResponse() @pytest.mark.parametrize('''urls_type''' , [str, list, dict] ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Any: '''simple docstring''' import requests monkeypatch.setattr(UpperCamelCase__ , '''request''' , UpperCamelCase__ ) UpperCAmelCase = URL if issubclass(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = url elif issubclass(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = [url] elif issubclass(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = {'''train''': url} UpperCAmelCase = '''dummy''' UpperCAmelCase = '''downloads''' UpperCAmelCase = tmp_path UpperCAmelCase = DownloadConfig( cache_dir=os.path.join(UpperCamelCase__ , UpperCamelCase__ ) , use_etag=UpperCamelCase__ , ) UpperCAmelCase = DownloadManager(dataset_name=UpperCamelCase__ , download_config=UpperCamelCase__ ) UpperCAmelCase = dl_manager.download(UpperCamelCase__ ) UpperCAmelCase = urls for downloaded_paths in [downloaded_paths]: if isinstance(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = [downloaded_paths] UpperCAmelCase = [urls] elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): assert "train" in downloaded_paths.keys() UpperCAmelCase = downloaded_paths.values() UpperCAmelCase = urls.values() assert downloaded_paths for downloaded_path, input_url in zip(UpperCamelCase__ , UpperCamelCase__ ): assert downloaded_path == dl_manager.downloaded_paths[input_url] UpperCAmelCase = Path(UpperCamelCase__ ) UpperCAmelCase = downloaded_path.parts assert parts[-1] == HASH assert parts[-2] == cache_subdir assert downloaded_path.exists() UpperCAmelCase = downloaded_path.read_text() assert content == CONTENT UpperCAmelCase = downloaded_path.with_suffix('''.json''' ) assert metadata_downloaded_path.exists() UpperCAmelCase = json.loads(metadata_downloaded_path.read_text() ) assert metadata_content == {"url": URL, "etag": None} @pytest.mark.parametrize('''paths_type''' , [str, list, dict] ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase = str(UpperCamelCase__ ) if issubclass(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = filename elif issubclass(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = [filename] elif issubclass(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = {'''train''': filename} UpperCAmelCase = '''dummy''' UpperCAmelCase = xz_file.parent UpperCAmelCase = '''extracted''' UpperCAmelCase = DownloadConfig( cache_dir=UpperCamelCase__ , use_etag=UpperCamelCase__ , ) UpperCAmelCase = DownloadManager(dataset_name=UpperCamelCase__ , download_config=UpperCamelCase__ ) UpperCAmelCase = dl_manager.extract(UpperCamelCase__ ) UpperCAmelCase = paths for extracted_paths in [extracted_paths]: if isinstance(UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase = [extracted_paths] UpperCAmelCase = [paths] elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): assert "train" in extracted_paths.keys() UpperCAmelCase = extracted_paths.values() UpperCAmelCase = paths.values() assert extracted_paths for extracted_path, input_path in zip(UpperCamelCase__ , UpperCamelCase__ ): assert extracted_path == dl_manager.extracted_paths[input_path] UpperCAmelCase = Path(UpperCamelCase__ ) UpperCAmelCase = extracted_path.parts assert parts[-1] == hash_url_to_filename(UpperCamelCase__ , etag=UpperCamelCase__ ) assert parts[-2] == extracted_subdir assert extracted_path.exists() UpperCAmelCase = extracted_path.read_text() UpperCAmelCase = text_file.read_text() assert extracted_file_content == expected_file_content def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' assert path.endswith('''.jsonl''' ) for num_items, line in enumerate(UpperCamelCase__ , start=1 ): UpperCAmelCase = json.loads(line.decode('''utf-8''' ) ) assert item.keys() == {"col_1", "col_2", "col_3"} assert num_items == 4 @pytest.mark.parametrize('''archive_jsonl''' , ['''tar_jsonl_path''', '''zip_jsonl_path'''] ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> Tuple: '''simple docstring''' UpperCAmelCase = request.getfixturevalue(UpperCamelCase__ ) UpperCAmelCase = DownloadManager() for num_jsonl, (path, file) in enumerate(dl_manager.iter_archive(UpperCamelCase__ ) , start=1 ): _test_jsonl(UpperCamelCase__ , UpperCamelCase__ ) assert num_jsonl == 2 @pytest.mark.parametrize('''archive_nested_jsonl''' , ['''tar_nested_jsonl_path''', '''zip_nested_jsonl_path'''] ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' UpperCAmelCase = request.getfixturevalue(UpperCamelCase__ ) UpperCAmelCase = DownloadManager() for num_tar, (path, file) in enumerate(dl_manager.iter_archive(UpperCamelCase__ ) , start=1 ): for num_jsonl, (subpath, subfile) in enumerate(dl_manager.iter_archive(UpperCamelCase__ ) , start=1 ): _test_jsonl(UpperCamelCase__ , UpperCamelCase__ ) assert num_tar == 1 assert num_jsonl == 2 def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> int: '''simple docstring''' UpperCAmelCase = DownloadManager() for num_file, file in enumerate(dl_manager.iter_files(UpperCamelCase__ ) , start=1 ): assert os.path.basename(UpperCamelCase__ ) == ("test.txt" if num_file == 1 else "train.txt") assert num_file == 2
273
'''simple docstring''' def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = len(lowerCAmelCase ) for i in range(length - 1 ): _lowerCAmelCase = i for k in range(i + 1 , lowerCAmelCase ): if collection[k] < collection[least]: _lowerCAmelCase = k if least != i: _lowerCAmelCase , _lowerCAmelCase = (collection[i], collection[least]) return collection if __name__ == "__main__": A__ : str =input('''Enter numbers separated by a comma:\n''').strip() A__ : Optional[int] =[int(item) for item in user_input.split(''',''')] print(selection_sort(unsorted))
70
0
'''simple docstring''' import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () _UpperCamelCase = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). _UpperCamelCase = [0, 25, 50] _UpperCamelCase = [25, 50, 75] _UpperCamelCase = fuzz.membership.trimf(X, abca) _UpperCamelCase = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. _UpperCamelCase = np.ones(75) _UpperCamelCase = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) _UpperCamelCase = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) _UpperCamelCase = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) _UpperCamelCase = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) _UpperCamelCase = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] _UpperCamelCase = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) _UpperCamelCase = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] _UpperCamelCase = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] _UpperCamelCase = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
208
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL A__ : List[str] =logging.get_logger(__name__) def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if isinstance(lowerCAmelCase , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowerCAmelCase , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowerCAmelCase ): return [[videos]] raise ValueError(f"Could not make batched video from {videos}" ) class UpperCAmelCase ( snake_case_ ): _lowercase: Any = ['''pixel_values'''] def __init__( self : Tuple , __snake_case : bool = True , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = PILImageResampling.BILINEAR , __snake_case : bool = True , __snake_case : Dict[str, int] = None , __snake_case : bool = True , __snake_case : Union[int, float] = 1 / 2_55 , __snake_case : bool = True , __snake_case : bool = True , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , **__snake_case : str , ) -> None: super().__init__(**__snake_case ) _lowerCAmelCase = size if size is not None else {"""shortest_edge""": 2_56} _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) _lowerCAmelCase = crop_size if crop_size is not None else {"""height""": 2_24, """width""": 2_24} _lowerCAmelCase = get_size_dict(__snake_case , param_name="""crop_size""" ) _lowerCAmelCase = do_resize _lowerCAmelCase = size _lowerCAmelCase = do_center_crop _lowerCAmelCase = crop_size _lowerCAmelCase = resample _lowerCAmelCase = do_rescale _lowerCAmelCase = rescale_factor _lowerCAmelCase = offset _lowerCAmelCase = do_normalize _lowerCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase__ ( self : int , __snake_case : np.ndarray , __snake_case : Dict[str, int] , __snake_case : PILImageResampling = PILImageResampling.BILINEAR , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Optional[Any] , ) -> np.ndarray: _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) if "shortest_edge" in size: _lowerCAmelCase = get_resize_output_image_size(__snake_case , size["""shortest_edge"""] , default_to_square=__snake_case ) elif "height" in size and "width" in size: _lowerCAmelCase = (size["""height"""], size["""width"""]) else: raise ValueError(f"Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}" ) return resize(__snake_case , size=__snake_case , resample=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Union[str, Any] , __snake_case : np.ndarray , __snake_case : Dict[str, int] , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : List[Any] , ) -> np.ndarray: _lowerCAmelCase = get_size_dict(__snake_case ) if "height" not in size or "width" not in size: raise ValueError(f"Size must have 'height' and 'width' as keys. Got {size.keys()}" ) return center_crop(__snake_case , size=(size["""height"""], size["""width"""]) , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Union[str, Any] , __snake_case : np.ndarray , __snake_case : Union[int, float] , __snake_case : bool = True , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Optional[Any] , ) -> Dict: _lowerCAmelCase = image.astype(np.floataa ) if offset: _lowerCAmelCase = image - (scale / 2) return rescale(__snake_case , scale=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Optional[int] , __snake_case : np.ndarray , __snake_case : Union[float, List[float]] , __snake_case : Union[float, List[float]] , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Tuple , ) -> np.ndarray: return normalize(__snake_case , mean=__snake_case , std=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : List[Any] , __snake_case : ImageInput , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = None , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : bool = None , __snake_case : float = None , __snake_case : bool = None , __snake_case : bool = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) if offset and not do_rescale: raise ValueError("""For offset, do_rescale must also be set to True.""" ) # All transformations expect numpy arrays. _lowerCAmelCase = to_numpy_array(__snake_case ) if do_resize: _lowerCAmelCase = self.resize(image=__snake_case , size=__snake_case , resample=__snake_case ) if do_center_crop: _lowerCAmelCase = self.center_crop(__snake_case , size=__snake_case ) if do_rescale: _lowerCAmelCase = self.rescale(image=__snake_case , scale=__snake_case , offset=__snake_case ) if do_normalize: _lowerCAmelCase = self.normalize(image=__snake_case , mean=__snake_case , std=__snake_case ) _lowerCAmelCase = to_channel_dimension_format(__snake_case , __snake_case ) return image def lowercase__ ( self : List[Any] , __snake_case : ImageInput , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = None , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : bool = None , __snake_case : float = None , __snake_case : bool = None , __snake_case : bool = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[str, TensorType]] = None , __snake_case : ChannelDimension = ChannelDimension.FIRST , **__snake_case : List[str] , ) -> PIL.Image.Image: _lowerCAmelCase = do_resize if do_resize is not None else self.do_resize _lowerCAmelCase = resample if resample is not None else self.resample _lowerCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop _lowerCAmelCase = do_rescale if do_rescale is not None else self.do_rescale _lowerCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCAmelCase = offset if offset is not None else self.offset _lowerCAmelCase = do_normalize if do_normalize is not None else self.do_normalize _lowerCAmelCase = image_mean if image_mean is not None else self.image_mean _lowerCAmelCase = image_std if image_std is not None else self.image_std _lowerCAmelCase = size if size is not None else self.size _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) _lowerCAmelCase = crop_size if crop_size is not None else self.crop_size _lowerCAmelCase = get_size_dict(__snake_case , param_name="""crop_size""" ) if not valid_images(__snake_case ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) _lowerCAmelCase = make_batched(__snake_case ) _lowerCAmelCase = [ [ self._preprocess_image( image=__snake_case , do_resize=__snake_case , size=__snake_case , resample=__snake_case , do_center_crop=__snake_case , crop_size=__snake_case , do_rescale=__snake_case , rescale_factor=__snake_case , offset=__snake_case , do_normalize=__snake_case , image_mean=__snake_case , image_std=__snake_case , data_format=__snake_case , ) for img in video ] for video in videos ] _lowerCAmelCase = {"""pixel_values""": videos} return BatchFeature(data=__snake_case , tensor_type=__snake_case )
70
0
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { '''microsoft/git-base''': '''https://huggingface.co/microsoft/git-base/resolve/main/config.json''', } class lowercase ( snake_case_ ): """simple docstring""" _a = '''git_vision_model''' def __init__( self , UpperCamelCase_=768 , UpperCamelCase_=3072 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=3 , UpperCamelCase_=224 , UpperCamelCase_=16 , UpperCamelCase_="quick_gelu" , UpperCamelCase_=1e-5 , UpperCamelCase_=0.0 , UpperCamelCase_=0.02 , **UpperCamelCase_ , ): '''simple docstring''' super().__init__(**__snake_case ) UpperCamelCase__ :Union[str, Any] = hidden_size UpperCamelCase__ :Optional[Any] = intermediate_size UpperCamelCase__ :List[Any] = num_hidden_layers UpperCamelCase__ :List[str] = num_attention_heads UpperCamelCase__ :Optional[Any] = num_channels UpperCamelCase__ :List[Any] = patch_size UpperCamelCase__ :Dict = image_size UpperCamelCase__ :str = initializer_range UpperCamelCase__ :Optional[Any] = attention_dropout UpperCamelCase__ :Dict = layer_norm_eps UpperCamelCase__ :List[str] = hidden_act @classmethod def lowerCAmelCase__ ( cls , UpperCamelCase_ , **UpperCamelCase_ ): '''simple docstring''' cls._set_token_in_kwargs(__snake_case ) UpperCamelCase__ , UpperCamelCase__ :Optional[int] = cls.get_config_dict(__snake_case , **__snake_case ) # get the vision config dict if we are loading from GITConfig if config_dict.get('''model_type''' ) == "git": UpperCamelCase__ :List[Any] = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__snake_case , **__snake_case ) class lowercase ( snake_case_ ): """simple docstring""" _a = '''git''' def __init__( self , UpperCamelCase_=None , UpperCamelCase_=30522 , UpperCamelCase_=768 , UpperCamelCase_=6 , UpperCamelCase_=12 , UpperCamelCase_=3072 , UpperCamelCase_="gelu" , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=1024 , UpperCamelCase_=0.02 , UpperCamelCase_=1e-12 , UpperCamelCase_=0 , UpperCamelCase_="absolute" , UpperCamelCase_=True , UpperCamelCase_=False , UpperCamelCase_=101 , UpperCamelCase_=102 , UpperCamelCase_=None , **UpperCamelCase_ , ): '''simple docstring''' super().__init__(bos_token_id=__snake_case , eos_token_id=__snake_case , pad_token_id=__snake_case , **__snake_case ) if vision_config is None: UpperCamelCase__ :List[Any] = {} logger.info('''vision_config is None. initializing the GitVisionConfig with default values.''' ) UpperCamelCase__ :int = GitVisionConfig(**__snake_case ) UpperCamelCase__ :List[str] = vocab_size UpperCamelCase__ :Optional[Any] = hidden_size UpperCamelCase__ :str = num_hidden_layers UpperCamelCase__ :List[Any] = num_attention_heads UpperCamelCase__ :Dict = hidden_act UpperCamelCase__ :Tuple = intermediate_size UpperCamelCase__ :Optional[int] = hidden_dropout_prob UpperCamelCase__ :int = attention_probs_dropout_prob UpperCamelCase__ :List[str] = max_position_embeddings UpperCamelCase__ :List[Any] = initializer_range UpperCamelCase__ :Union[str, Any] = layer_norm_eps UpperCamelCase__ :int = position_embedding_type UpperCamelCase__ :Optional[int] = use_cache UpperCamelCase__ :Optional[int] = tie_word_embeddings UpperCamelCase__ :Dict = num_image_with_embedding UpperCamelCase__ :Union[str, Any] = bos_token_id UpperCamelCase__ :int = eos_token_id def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = copy.deepcopy(self.__dict__ ) UpperCamelCase__ :Union[str, Any] = self.vision_config.to_dict() UpperCamelCase__ :int = self.__class__.model_type return output
97
'''simple docstring''' import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class UpperCAmelCase ( snake_case_ ): _lowercase: Union[str, Any] = ['''image_processor''', '''tokenizer'''] _lowercase: int = '''AutoImageProcessor''' _lowercase: Optional[int] = '''AutoTokenizer''' def __init__( self : int , __snake_case : Tuple=None , __snake_case : Optional[int]=None , **__snake_case : Tuple ) -> List[Any]: _lowerCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __snake_case , ) _lowerCAmelCase = kwargs.pop("""feature_extractor""" ) _lowerCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__snake_case , __snake_case ) _lowerCAmelCase = self.image_processor _lowerCAmelCase = False def __call__( self : Dict , *__snake_case : Optional[int] , **__snake_case : Union[str, Any] ) -> Tuple: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*__snake_case , **__snake_case ) _lowerCAmelCase = kwargs.pop("""images""" , __snake_case ) _lowerCAmelCase = kwargs.pop("""text""" , __snake_case ) if len(__snake_case ) > 0: _lowerCAmelCase = args[0] _lowerCAmelCase = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _lowerCAmelCase = self.image_processor(__snake_case , *__snake_case , **__snake_case ) if text is not None: _lowerCAmelCase = self.tokenizer(__snake_case , **__snake_case ) if text is None: return inputs elif images is None: return encodings else: _lowerCAmelCase = encodings["""input_ids"""] return inputs def lowercase__ ( self : List[Any] , *__snake_case : Dict , **__snake_case : List[str] ) -> int: return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def lowercase__ ( self : int , *__snake_case : Tuple , **__snake_case : Optional[Any] ) -> Any: return self.tokenizer.decode(*__snake_case , **__snake_case ) @contextmanager def lowercase__ ( self : int ) -> Optional[Any]: warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _lowerCAmelCase = True _lowerCAmelCase = self.tokenizer yield _lowerCAmelCase = self.image_processor _lowerCAmelCase = False def lowercase__ ( self : Optional[int] , __snake_case : Union[str, Any] , __snake_case : List[Any]=False , __snake_case : Dict=None ) -> Tuple: if added_vocab is None: _lowerCAmelCase = self.tokenizer.get_added_vocab() _lowerCAmelCase = {} while tokens: _lowerCAmelCase = re.search(R"""<s_(.*?)>""" , __snake_case , re.IGNORECASE ) if start_token is None: break _lowerCAmelCase = start_token.group(1 ) _lowerCAmelCase = re.search(Rf"</s_{key}>" , __snake_case , re.IGNORECASE ) _lowerCAmelCase = start_token.group() if end_token is None: _lowerCAmelCase = tokens.replace(__snake_case , """""" ) else: _lowerCAmelCase = end_token.group() _lowerCAmelCase = re.escape(__snake_case ) _lowerCAmelCase = re.escape(__snake_case ) _lowerCAmelCase = re.search(f"{start_token_escaped}(.*?){end_token_escaped}" , __snake_case , re.IGNORECASE ) if content is not None: _lowerCAmelCase = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _lowerCAmelCase = self.tokenajson(__snake_case , is_inner_value=__snake_case , added_vocab=__snake_case ) if value: if len(__snake_case ) == 1: _lowerCAmelCase = value[0] _lowerCAmelCase = value else: # leaf nodes _lowerCAmelCase = [] for leaf in content.split(R"""<sep/>""" ): _lowerCAmelCase = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _lowerCAmelCase = leaf[1:-2] # for categorical special tokens output[key].append(__snake_case ) if len(output[key] ) == 1: _lowerCAmelCase = output[key][0] _lowerCAmelCase = tokens[tokens.find(__snake_case ) + len(__snake_case ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=__snake_case , added_vocab=__snake_case ) if len(__snake_case ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def lowercase__ ( self : Optional[int] ) -> Union[str, Any]: warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , __snake_case , ) return self.image_processor_class @property def lowercase__ ( self : List[Any] ) -> Any: warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , __snake_case , ) return self.image_processor
70
0
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin lowerCamelCase_ : Dict = ''' Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] ''' class __A ( unittest.TestCase, snake_case_ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: a =load_tool('''text-question-answering''' ) self.tool.setup() a =load_tool('''text-question-answering''' , remote=__snake_case ) def SCREAMING_SNAKE_CASE ( self ) -> int: a =self.tool(__snake_case , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(__snake_case , '''launched the BigScience Research Workshop''' ) def SCREAMING_SNAKE_CASE ( self ) -> Any: a =self.remote_tool(__snake_case , '''What did Hugging Face do in April 2021?''' ) self.assertEqual(__snake_case , '''launched the BigScience Research Workshop''' ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: a =self.tool(text=__snake_case , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(__snake_case , '''launched the BigScience Research Workshop''' ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: a =self.remote_tool(text=__snake_case , question='''What did Hugging Face do in April 2021?''' ) self.assertEqual(__snake_case , '''launched the BigScience Research Workshop''' )
81
'''simple docstring''' from __future__ import annotations import math def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True A__ : Optional[Any] =[num for num in range(3, 10_00_01, 2) if not is_prime(num)] def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if not isinstance(lowerCAmelCase , lowerCAmelCase ): raise ValueError("""n must be an integer""" ) if n <= 0: raise ValueError("""n must be >= 0""" ) _lowerCAmelCase = [] for num in range(len(lowerCAmelCase ) ): _lowerCAmelCase = 0 while 2 * i * i <= odd_composites[num]: _lowerCAmelCase = odd_composites[num] - 2 * i * i if is_prime(lowerCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(lowerCAmelCase ) == n: return list_nums return [] def UpperCamelCase__ ( ): """simple docstring""" return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
70
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL snake_case_ = logging.get_logger(__name__) def _lowerCAmelCase ( lowercase_ ): if isinstance(lowercase_ , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowercase_ , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowercase_ ): return [[videos]] raise ValueError(F"""Could not make batched video from {videos}""" ) class A_ ( snake_case_ ): """simple docstring""" __UpperCamelCase = ['''pixel_values'''] def __init__( self :Tuple , lowercase_ :bool = True , lowercase_ :Dict[str, int] = None , lowercase_ :PILImageResampling = PILImageResampling.BILINEAR , lowercase_ :bool = True , lowercase_ :Dict[str, int] = None , lowercase_ :bool = True , lowercase_ :Union[int, float] = 1 / 2_55 , lowercase_ :bool = True , lowercase_ :bool = True , lowercase_ :Optional[Union[float, List[float]]] = None , lowercase_ :Optional[Union[float, List[float]]] = None , **lowercase_ :str , ) -> None: super().__init__(**__snake_case ) UpperCAmelCase = size if size is not None else {'shortest_edge': 2_56} UpperCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) UpperCAmelCase = crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} UpperCAmelCase = get_size_dict(__snake_case , param_name='crop_size' ) UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = resample UpperCAmelCase = do_rescale UpperCAmelCase = rescale_factor UpperCAmelCase = offset UpperCAmelCase = do_normalize UpperCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__ ( self :int , lowercase_ :np.ndarray , lowercase_ :Dict[str, int] , lowercase_ :PILImageResampling = PILImageResampling.BILINEAR , lowercase_ :Optional[Union[str, ChannelDimension]] = None , **lowercase_ :Optional[Any] , ) -> np.ndarray: UpperCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) if "shortest_edge" in size: UpperCAmelCase = get_resize_output_image_size(__snake_case , size['shortest_edge'] , default_to_square=__snake_case ) elif "height" in size and "width" in size: UpperCAmelCase = (size['height'], size['width']) else: raise ValueError(f"""Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}""" ) return resize(__snake_case , size=__snake_case , resample=__snake_case , data_format=__snake_case , **__snake_case ) def UpperCAmelCase__ ( self :Union[str, Any] , lowercase_ :np.ndarray , lowercase_ :Dict[str, int] , lowercase_ :Optional[Union[str, ChannelDimension]] = None , **lowercase_ :List[Any] , ) -> np.ndarray: UpperCAmelCase = get_size_dict(__snake_case ) if "height" not in size or "width" not in size: raise ValueError(f"""Size must have 'height' and 'width' as keys. Got {size.keys()}""" ) return center_crop(__snake_case , size=(size['height'], size['width']) , data_format=__snake_case , **__snake_case ) def UpperCAmelCase__ ( self :Union[str, Any] , lowercase_ :np.ndarray , lowercase_ :Union[int, float] , lowercase_ :bool = True , lowercase_ :Optional[Union[str, ChannelDimension]] = None , **lowercase_ :Optional[Any] , ) -> Dict: UpperCAmelCase = image.astype(np.floataa ) if offset: UpperCAmelCase = image - (scale / 2) return rescale(__snake_case , scale=__snake_case , data_format=__snake_case , **__snake_case ) def UpperCAmelCase__ ( self :Optional[int] , lowercase_ :np.ndarray , lowercase_ :Union[float, List[float]] , lowercase_ :Union[float, List[float]] , lowercase_ :Optional[Union[str, ChannelDimension]] = None , **lowercase_ :Tuple , ) -> np.ndarray: return normalize(__snake_case , mean=__snake_case , std=__snake_case , data_format=__snake_case , **__snake_case ) def UpperCAmelCase__ ( self :List[Any] , lowercase_ :ImageInput , lowercase_ :bool = None , lowercase_ :Dict[str, int] = None , lowercase_ :PILImageResampling = None , lowercase_ :bool = None , lowercase_ :Dict[str, int] = None , lowercase_ :bool = None , lowercase_ :float = None , lowercase_ :bool = None , lowercase_ :bool = None , lowercase_ :Optional[Union[float, List[float]]] = None , lowercase_ :Optional[Union[float, List[float]]] = None , lowercase_ :Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) if offset and not do_rescale: raise ValueError('For offset, do_rescale must also be set to True.' ) # All transformations expect numpy arrays. UpperCAmelCase = to_numpy_array(__snake_case ) if do_resize: UpperCAmelCase = self.resize(image=__snake_case , size=__snake_case , resample=__snake_case ) if do_center_crop: UpperCAmelCase = self.center_crop(__snake_case , size=__snake_case ) if do_rescale: UpperCAmelCase = self.rescale(image=__snake_case , scale=__snake_case , offset=__snake_case ) if do_normalize: UpperCAmelCase = self.normalize(image=__snake_case , mean=__snake_case , std=__snake_case ) UpperCAmelCase = to_channel_dimension_format(__snake_case , __snake_case ) return image def UpperCAmelCase__ ( self :List[Any] , lowercase_ :ImageInput , lowercase_ :bool = None , lowercase_ :Dict[str, int] = None , lowercase_ :PILImageResampling = None , lowercase_ :bool = None , lowercase_ :Dict[str, int] = None , lowercase_ :bool = None , lowercase_ :float = None , lowercase_ :bool = None , lowercase_ :bool = None , lowercase_ :Optional[Union[float, List[float]]] = None , lowercase_ :Optional[Union[float, List[float]]] = None , lowercase_ :Optional[Union[str, TensorType]] = None , lowercase_ :ChannelDimension = ChannelDimension.FIRST , **lowercase_ :List[str] , ) -> PIL.Image.Image: UpperCAmelCase = do_resize if do_resize is not None else self.do_resize UpperCAmelCase = resample if resample is not None else self.resample UpperCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase = offset if offset is not None else self.offset UpperCAmelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase = image_mean if image_mean is not None else self.image_mean UpperCAmelCase = image_std if image_std is not None else self.image_std UpperCAmelCase = size if size is not None else self.size UpperCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) UpperCAmelCase = crop_size if crop_size is not None else self.crop_size UpperCAmelCase = get_size_dict(__snake_case , param_name='crop_size' ) if not valid_images(__snake_case ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) UpperCAmelCase = make_batched(__snake_case ) UpperCAmelCase = [ [ self._preprocess_image( image=__snake_case , do_resize=__snake_case , size=__snake_case , resample=__snake_case , do_center_crop=__snake_case , crop_size=__snake_case , do_rescale=__snake_case , rescale_factor=__snake_case , offset=__snake_case , do_normalize=__snake_case , image_mean=__snake_case , image_std=__snake_case , data_format=__snake_case , ) for img in video ] for video in videos ] UpperCAmelCase = {'pixel_values': videos} return BatchFeature(data=__snake_case , tensor_type=__snake_case )
78
'''simple docstring''' import argparse import json from tqdm import tqdm def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--src_path""" , type=lowerCAmelCase , default="""biencoder-nq-dev.json""" , help="""Path to raw DPR training data""" , ) parser.add_argument( """--evaluation_set""" , type=lowerCAmelCase , help="""where to store parsed evaluation_set file""" , ) parser.add_argument( """--gold_data_path""" , type=lowerCAmelCase , help="""where to store parsed gold_data_path file""" , ) _lowerCAmelCase = parser.parse_args() with open(args.src_path , """r""" ) as src_file, open(args.evaluation_set , """w""" ) as eval_file, open( args.gold_data_path , """w""" ) as gold_file: _lowerCAmelCase = json.load(lowerCAmelCase ) for dpr_record in tqdm(lowerCAmelCase ): _lowerCAmelCase = dpr_record["""question"""] _lowerCAmelCase = [context["""title"""] for context in dpr_record["""positive_ctxs"""]] eval_file.write(question + """\n""" ) gold_file.write("""\t""".join(lowerCAmelCase ) + """\n""" ) if __name__ == "__main__": main()
70
0
"""simple docstring""" import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel __SCREAMING_SNAKE_CASE =False __SCREAMING_SNAKE_CASE =True __SCREAMING_SNAKE_CASE =False if __name__ == "__main__": __SCREAMING_SNAKE_CASE =argparse.ArgumentParser() parser.add_argument( "--repo_path", default=None, type=str, required=True, help="The config json file corresponding to the architecture.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") __SCREAMING_SNAKE_CASE =parser.parse_args() __SCREAMING_SNAKE_CASE ={ '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } __SCREAMING_SNAKE_CASE ={ '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } __SCREAMING_SNAKE_CASE ='''''' if has_file(args.repo_path, "config.json") else '''unet''' with open(os.path.join(args.repo_path, subfolder, "config.json"), "r", encoding="utf-8") as reader: __SCREAMING_SNAKE_CASE =reader.read() __SCREAMING_SNAKE_CASE =json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, "config.json"): __SCREAMING_SNAKE_CASE =UNetaDModel(**config) else: __SCREAMING_SNAKE_CASE =UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel __SCREAMING_SNAKE_CASE =class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) __SCREAMING_SNAKE_CASE =dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: __SCREAMING_SNAKE_CASE =config[key] del config[key] __SCREAMING_SNAKE_CASE =[k.replace("UNetRes", "") for k in config['''down_block_types''']] __SCREAMING_SNAKE_CASE =[k.replace("UNetRes", "") for k in config['''up_block_types''']] if do_only_weights: __SCREAMING_SNAKE_CASE =torch.load(os.path.join(args.repo_path, subfolder, "diffusion_pytorch_model.bin")) __SCREAMING_SNAKE_CASE ={} for param_key, param_value in state_dict.items(): if param_key.endswith(".op.bias") or param_key.endswith(".op.weight"): continue __SCREAMING_SNAKE_CASE =False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(".")[0] == key: __SCREAMING_SNAKE_CASE =param_value __SCREAMING_SNAKE_CASE =True if not has_changed: __SCREAMING_SNAKE_CASE =param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
213
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : Optional[Any] ={ '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Union[str, Any] =[ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys A__ : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
70
0
SCREAMING_SNAKE_CASE_ = [sum(int(c, 1_0) ** 2 for c in i.__str__()) for i in range(1_0_0_0_0_0)] def __lowercase ( _SCREAMING_SNAKE_CASE ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_00_00] number //= 10_00_00 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution SCREAMING_SNAKE_CASE_ = [None] * 1_0_0_0_0_0_0_0 SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = False def __lowercase ( _SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore SCREAMING_SNAKE_CASE = chain(next_number(_SCREAMING_SNAKE_CASE ) ) SCREAMING_SNAKE_CASE = number_chain while number < 10_00_00_00: SCREAMING_SNAKE_CASE = number_chain number *= 10 return number_chain def __lowercase ( _SCREAMING_SNAKE_CASE = 10_00_00_00 ) -> Optional[int]: '''simple docstring''' for i in range(1 , _SCREAMING_SNAKE_CASE ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
296
'''simple docstring''' import math from typing import Callable, List, Optional, Union import numpy as np import PIL import torch from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from diffusers.schedulers import DDIMScheduler, DDPMScheduler, LMSDiscreteScheduler, PNDMScheduler def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=[] ): """simple docstring""" _lowerCAmelCase = size[0] - overlap_pixels * 2 _lowerCAmelCase = size[1] - overlap_pixels * 2 for letter in ["l", "r"]: if letter in remove_borders: size_x += overlap_pixels for letter in ["t", "b"]: if letter in remove_borders: size_y += overlap_pixels _lowerCAmelCase = np.ones((size_y, size_x) , dtype=np.uinta ) * 2_55 _lowerCAmelCase = np.pad(lowerCAmelCase , mode="""linear_ramp""" , pad_width=lowerCAmelCase , end_values=0 ) if "l" in remove_borders: _lowerCAmelCase = mask[:, overlap_pixels : mask.shape[1]] if "r" in remove_borders: _lowerCAmelCase = mask[:, 0 : mask.shape[1] - overlap_pixels] if "t" in remove_borders: _lowerCAmelCase = mask[overlap_pixels : mask.shape[0], :] if "b" in remove_borders: _lowerCAmelCase = mask[0 : mask.shape[0] - overlap_pixels, :] return mask def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" return max(lowerCAmelCase , min(lowerCAmelCase , lowerCAmelCase ) ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" return ( clamp(rect[0] , min[0] , max[0] ), clamp(rect[1] , min[1] , max[1] ), clamp(rect[2] , min[0] , max[0] ), clamp(rect[3] , min[1] , max[1] ), ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = list(lowerCAmelCase ) rect[0] -= overlap rect[1] -= overlap rect[2] += overlap rect[3] += overlap _lowerCAmelCase = clamp_rect(lowerCAmelCase , [0, 0] , [image_size[0], image_size[1]] ) return rect def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = Image.new("""RGB""" , (tile.size[0] + original_slice, tile.size[1]) ) result.paste( original_image.resize((tile.size[0], tile.size[1]) , Image.BICUBIC ).crop( (slice_x, 0, slice_x + original_slice, tile.size[1]) ) , (0, 0) , ) result.paste(lowerCAmelCase , (original_slice, 0) ) return result def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = (original_image_slice * 4, 0, tile.size[0], tile.size[1]) _lowerCAmelCase = tile.crop(lowerCAmelCase ) return tile def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = n % d return n - divisor class UpperCAmelCase ( snake_case_ ): def __init__( self : List[Any] , __snake_case : AutoencoderKL , __snake_case : CLIPTextModel , __snake_case : CLIPTokenizer , __snake_case : UNetaDConditionModel , __snake_case : DDPMScheduler , __snake_case : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , __snake_case : int = 3_50 , ) -> int: super().__init__( vae=__snake_case , text_encoder=__snake_case , tokenizer=__snake_case , unet=__snake_case , low_res_scheduler=__snake_case , scheduler=__snake_case , max_noise_level=__snake_case , ) def lowercase__ ( self : List[Any] , __snake_case : List[Any] , __snake_case : List[Any] , __snake_case : Dict , __snake_case : List[Any] , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Any , **__snake_case : str ) -> int: torch.manual_seed(0 ) _lowerCAmelCase = ( min(image.size[0] - (tile_size + original_image_slice) , x * tile_size ), min(image.size[1] - (tile_size + original_image_slice) , y * tile_size ), min(image.size[0] , (x + 1) * tile_size ), min(image.size[1] , (y + 1) * tile_size ), ) _lowerCAmelCase = add_overlap_rect(__snake_case , __snake_case , image.size ) _lowerCAmelCase = image.crop(__snake_case ) _lowerCAmelCase = ((crop_rect[0] + ((crop_rect[2] - crop_rect[0]) / 2)) / image.size[0]) * tile.size[0] _lowerCAmelCase = translated_slice_x - (original_image_slice / 2) _lowerCAmelCase = max(0 , __snake_case ) _lowerCAmelCase = squeeze_tile(__snake_case , __snake_case , __snake_case , __snake_case ) _lowerCAmelCase = to_input.size _lowerCAmelCase = to_input.resize((tile_size, tile_size) , Image.BICUBIC ) _lowerCAmelCase = super(__snake_case , self ).__call__(image=__snake_case , **__snake_case ).images[0] _lowerCAmelCase = upscaled_tile.resize((orig_input_size[0] * 4, orig_input_size[1] * 4) , Image.BICUBIC ) _lowerCAmelCase = unsqueeze_tile(__snake_case , __snake_case ) _lowerCAmelCase = upscaled_tile.resize((tile.size[0] * 4, tile.size[1] * 4) , Image.BICUBIC ) _lowerCAmelCase = [] if x == 0: remove_borders.append("""l""" ) elif crop_rect[2] == image.size[0]: remove_borders.append("""r""" ) if y == 0: remove_borders.append("""t""" ) elif crop_rect[3] == image.size[1]: remove_borders.append("""b""" ) _lowerCAmelCase = Image.fromarray( make_transparency_mask( (upscaled_tile.size[0], upscaled_tile.size[1]) , tile_border * 4 , remove_borders=__snake_case ) , mode="""L""" , ) final_image.paste( __snake_case , (crop_rect_with_overlap[0] * 4, crop_rect_with_overlap[1] * 4) , __snake_case ) @torch.no_grad() def __call__( self : Union[str, Any] , __snake_case : Union[str, List[str]] , __snake_case : Union[PIL.Image.Image, List[PIL.Image.Image]] , __snake_case : int = 75 , __snake_case : float = 9.0 , __snake_case : int = 50 , __snake_case : Optional[Union[str, List[str]]] = None , __snake_case : Optional[int] = 1 , __snake_case : float = 0.0 , __snake_case : Optional[torch.Generator] = None , __snake_case : Optional[torch.FloatTensor] = None , __snake_case : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __snake_case : int = 1 , __snake_case : int = 1_28 , __snake_case : int = 32 , __snake_case : int = 32 , ) -> str: _lowerCAmelCase = Image.new("""RGB""" , (image.size[0] * 4, image.size[1] * 4) ) _lowerCAmelCase = math.ceil(image.size[0] / tile_size ) _lowerCAmelCase = math.ceil(image.size[1] / tile_size ) _lowerCAmelCase = tcx * tcy _lowerCAmelCase = 0 for y in range(__snake_case ): for x in range(__snake_case ): self._process_tile( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , prompt=__snake_case , num_inference_steps=__snake_case , guidance_scale=__snake_case , noise_level=__snake_case , negative_prompt=__snake_case , num_images_per_prompt=__snake_case , eta=__snake_case , generator=__snake_case , latents=__snake_case , ) current_count += 1 if callback is not None: callback({"""progress""": current_count / total_tile_count, """image""": final_image} ) return final_image def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = """stabilityai/stable-diffusion-x4-upscaler""" _lowerCAmelCase = StableDiffusionTiledUpscalePipeline.from_pretrained(lowerCAmelCase , revision="""fp16""" , torch_dtype=torch.floataa ) _lowerCAmelCase = pipe.to("""cuda""" ) _lowerCAmelCase = Image.open("""../../docs/source/imgs/diffusers_library.jpg""" ) def callback(lowerCAmelCase ): print(f"progress: {obj['progress']:.4f}" ) obj["image"].save("""diffusers_library_progress.jpg""" ) _lowerCAmelCase = pipe(image=lowerCAmelCase , prompt="""Black font, white background, vector""" , noise_level=40 , callback=lowerCAmelCase ) final_image.save("""diffusers_library.jpg""" ) if __name__ == "__main__": main()
70
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _A = { '''configuration_mega''': ['''MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MegaConfig''', '''MegaOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''MEGA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MegaForCausalLM''', '''MegaForMaskedLM''', '''MegaForMultipleChoice''', '''MegaForQuestionAnswering''', '''MegaForSequenceClassification''', '''MegaForTokenClassification''', '''MegaModel''', '''MegaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
171
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class UpperCAmelCase ( snake_case_ , unittest.TestCase ): _lowercase: int = KandinskyVaaImgaImgPipeline _lowercase: List[str] = ['''image_embeds''', '''negative_image_embeds''', '''image'''] _lowercase: Optional[int] = [ '''image_embeds''', '''negative_image_embeds''', '''image''', ] _lowercase: Tuple = [ '''generator''', '''height''', '''width''', '''strength''', '''guidance_scale''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] _lowercase: List[str] = False @property def lowercase__ ( self : str ) -> List[str]: return 32 @property def lowercase__ ( self : Optional[int] ) -> List[Any]: return 32 @property def lowercase__ ( self : Tuple ) -> str: return self.time_input_dim @property def lowercase__ ( self : Any ) -> Optional[int]: return self.time_input_dim * 4 @property def lowercase__ ( self : int ) -> Optional[Any]: return 1_00 @property def lowercase__ ( self : int ) -> Dict: torch.manual_seed(0 ) _lowerCAmelCase = { """in_channels""": 4, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } _lowerCAmelCase = UNetaDConditionModel(**__snake_case ) return model @property def lowercase__ ( self : Union[str, Any] ) -> Tuple: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowercase__ ( self : Dict ) -> str: torch.manual_seed(0 ) _lowerCAmelCase = VQModel(**self.dummy_movq_kwargs ) return model def lowercase__ ( self : Optional[int] ) -> Optional[int]: _lowerCAmelCase = self.dummy_unet _lowerCAmelCase = self.dummy_movq _lowerCAmelCase = { """num_train_timesteps""": 10_00, """beta_schedule""": """linear""", """beta_start""": 0.0_00_85, """beta_end""": 0.0_12, """clip_sample""": False, """set_alpha_to_one""": False, """steps_offset""": 0, """prediction_type""": """epsilon""", """thresholding""": False, } _lowerCAmelCase = DDIMScheduler(**__snake_case ) _lowerCAmelCase = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def lowercase__ ( self : int , __snake_case : List[str] , __snake_case : List[Any]=0 ) -> Union[str, Any]: _lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __snake_case ) # create init_image _lowerCAmelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((2_56, 2_56) ) if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = { """image""": init_image, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 10, """guidance_scale""": 7.0, """strength""": 0.2, """output_type""": """np""", } return inputs def lowercase__ ( self : str ) -> Tuple: _lowerCAmelCase = """cpu""" _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) _lowerCAmelCase = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = pipe(**self.get_dummy_inputs(__snake_case ) ) _lowerCAmelCase = output.images _lowerCAmelCase = pipe( **self.get_dummy_inputs(__snake_case ) , return_dict=__snake_case , )[0] _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _lowerCAmelCase = np.array( [0.6_19_97_78, 0.63_98_44_06, 0.46_14_57_85, 0.62_94_49_84, 0.5_62_22_15, 0.47_30_61_32, 0.47_44_14_56, 0.4_60_76_06, 0.48_71_92_63] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): def lowercase__ ( self : Any ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : int ) -> Dict: _lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_img2img_frog.npy""" ) _lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) _lowerCAmelCase = """A red cartoon frog, 4k""" _lowerCAmelCase = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__snake_case ) _lowerCAmelCase = KandinskyVaaImgaImgPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-decoder""" , torch_dtype=torch.floataa ) _lowerCAmelCase = pipeline.to(__snake_case ) pipeline.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase , _lowerCAmelCase = pipe_prior( __snake_case , generator=__snake_case , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() _lowerCAmelCase = pipeline( image=__snake_case , image_embeds=__snake_case , negative_image_embeds=__snake_case , generator=__snake_case , num_inference_steps=1_00 , height=7_68 , width=7_68 , strength=0.2 , output_type="""np""" , ) _lowerCAmelCase = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(__snake_case , __snake_case )
70
0
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { '''EleutherAI/gpt-neo-1.3B''': '''https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json''', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class __lowercase (snake_case_ ): _UpperCamelCase = '''gpt_neo''' _UpperCamelCase = ['''past_key_values'''] _UpperCamelCase = {'''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers'''} def __init__( self , A_=5_0257 , A_=2048 , A_=2048 , A_=24 , A_=[[["global", "local"], 12]] , A_=16 , A_=None , A_=256 , A_="gelu_new" , A_=0.0 , A_=0.0 , A_=0.0 , A_=0.1 , A_=1e-5 , A_=0.02 , A_=True , A_=5_0256 , A_=5_0256 , **A_ , ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : str = vocab_size __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : Dict = hidden_size __lowerCAmelCase : int = num_layers __lowerCAmelCase : int = num_heads __lowerCAmelCase : Optional[int] = intermediate_size __lowerCAmelCase : str = window_size __lowerCAmelCase : Union[str, Any] = activation_function __lowerCAmelCase : Optional[Any] = resid_dropout __lowerCAmelCase : Any = embed_dropout __lowerCAmelCase : Optional[int] = attention_dropout __lowerCAmelCase : Any = classifier_dropout __lowerCAmelCase : Union[str, Any] = layer_norm_epsilon __lowerCAmelCase : Dict = initializer_range __lowerCAmelCase : Optional[Any] = use_cache __lowerCAmelCase : Tuple = bos_token_id __lowerCAmelCase : List[str] = eos_token_id __lowerCAmelCase : Tuple = attention_types __lowerCAmelCase : Optional[int] = self.expand_attention_types_params(__snake_case ) if len(self.attention_layers ) != self.num_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.attention_layers)` == `config.num_layers` ''' f"""but is `len(config.attention_layers) = {len(self.attention_layers )}`, """ f"""`config.num_layers = {self.num_layers}`. """ '''`config.attention_layers` is prepared using `config.attention_types`. ''' '''Please verify the value of `config.attention_types` argument.''' ) super().__init__(bos_token_id=__snake_case , eos_token_id=__snake_case , **__snake_case ) @staticmethod def UpperCamelCase__ ( A_ ) ->Any: '''simple docstring''' __lowerCAmelCase : Any = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): import torch __lowerCAmelCase : List[str] = input.size() __lowerCAmelCase : str = len(lowercase__ ) __lowerCAmelCase : int = shape[dimension] __lowerCAmelCase : Any = torch.arange(0 , lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = torch.div(sizedim - size , lowercase__ , rounding_mode='''floor''' ) + 1 __lowerCAmelCase : Optional[int] = torch.arange(lowercase__ ) + low_indices[:min_length][:, None] __lowerCAmelCase : Union[str, Any] = [slice(lowercase__ )] * rank __lowerCAmelCase : Any = indices __lowerCAmelCase : int = input[s] __lowerCAmelCase : int = list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): import torch __lowerCAmelCase : str = torch.arange(1 , lowercase__ ) __lowerCAmelCase : Optional[Any] = torch.remainder(lowercase__ , lowercase__ ) __lowerCAmelCase : Dict = remainders == 0 __lowerCAmelCase : Optional[Any] = candidates[divisor_indices] __lowerCAmelCase : Dict = torch.max(lowercase__ ) return largest_divisor, torch.div(lowercase__ , lowercase__ , rounding_mode='''floor''' ) class __lowercase (snake_case_ ): @property def UpperCamelCase__ ( self ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' __lowerCAmelCase : List[str] = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}} ) if self.use_past: self.fill_with_past_key_values_(__snake_case , direction='''inputs''' ) __lowerCAmelCase : int = {0: '''batch''', 1: '''past_sequence + sequence'''} else: __lowerCAmelCase : Tuple = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return self._config.num_heads def UpperCamelCase__ ( self , A_ , A_ = -1 , A_ = -1 , A_ = False , A_ = None , ) ->Mapping[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[int] = super(__snake_case , self ).generate_dummy_inputs( __snake_case , batch_size=__snake_case , seq_length=__snake_case , is_pair=__snake_case , framework=__snake_case ) # We need to order the input in the way they appears in the forward() __lowerCAmelCase : Optional[Any] = OrderedDict({'''input_ids''': common_inputs['''input_ids''']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch __lowerCAmelCase, __lowerCAmelCase : Optional[int] = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values __lowerCAmelCase : Any = seqlen + 2 __lowerCAmelCase : str = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __lowerCAmelCase : List[str] = [ (torch.zeros(__snake_case ), torch.zeros(__snake_case )) for _ in range(self.num_layers ) ] __lowerCAmelCase : int = common_inputs['''attention_mask'''] if self.use_past: __lowerCAmelCase : Optional[Any] = ordered_inputs['''attention_mask'''].dtype __lowerCAmelCase : Dict = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(__snake_case , __snake_case , dtype=__snake_case )] , dim=1 ) return ordered_inputs @property def UpperCamelCase__ ( self ) ->int: '''simple docstring''' return 13
275
'''simple docstring''' import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class UpperCAmelCase ( snake_case_ ): def lowercase__ ( self : List[Any] ) -> Union[str, Any]: _lowerCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__snake_case , """hidden_sizes""" ) ) self.parent.assertTrue(hasattr(__snake_case , """num_attention_heads""" ) ) self.parent.assertTrue(hasattr(__snake_case , """num_encoder_blocks""" ) ) class UpperCAmelCase : def __init__( self : Optional[int] , __snake_case : str , __snake_case : Dict=13 , __snake_case : str=64 , __snake_case : Dict=3 , __snake_case : Dict=4 , __snake_case : Tuple=[2, 2, 2, 2] , __snake_case : int=[8, 4, 2, 1] , __snake_case : List[str]=[16, 32, 64, 1_28] , __snake_case : Optional[Any]=[1, 4, 8, 16] , __snake_case : Dict=[1, 2, 4, 8] , __snake_case : Optional[Any]=True , __snake_case : List[str]=True , __snake_case : int="gelu" , __snake_case : Optional[Any]=0.1 , __snake_case : Any=0.1 , __snake_case : Tuple=0.02 , __snake_case : Union[str, Any]=3 , __snake_case : Tuple=None , ) -> List[str]: _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = image_size _lowerCAmelCase = num_channels _lowerCAmelCase = num_encoder_blocks _lowerCAmelCase = sr_ratios _lowerCAmelCase = depths _lowerCAmelCase = hidden_sizes _lowerCAmelCase = downsampling_rates _lowerCAmelCase = num_attention_heads _lowerCAmelCase = is_training _lowerCAmelCase = use_labels _lowerCAmelCase = hidden_act _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = initializer_range _lowerCAmelCase = num_labels _lowerCAmelCase = scope def lowercase__ ( self : int ) -> Union[str, Any]: _lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase = None if self.use_labels: _lowerCAmelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _lowerCAmelCase = self.get_config() return config, pixel_values, labels def lowercase__ ( self : List[Any] ) -> List[str]: return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def lowercase__ ( self : Tuple , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : Optional[int] ) -> Tuple: _lowerCAmelCase = SegformerModel(config=__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = _lowerCAmelCase = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def lowercase__ ( self : List[str] , __snake_case : List[Any] , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> List[str]: _lowerCAmelCase = self.num_labels _lowerCAmelCase = SegformerForSemanticSegmentation(__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = model(__snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) _lowerCAmelCase = model(__snake_case , labels=__snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def lowercase__ ( self : str , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Dict ) -> List[str]: _lowerCAmelCase = 1 _lowerCAmelCase = SegformerForSemanticSegmentation(config=__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(__snake_case ) _lowerCAmelCase = model(__snake_case , labels=__snake_case ) self.parent.assertGreater(result.loss , 0.0 ) def lowercase__ ( self : Optional[int] ) -> int: _lowerCAmelCase = self.prepare_config_and_inputs() _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = config_and_inputs _lowerCAmelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Any = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) _lowercase: Tuple = ( { '''feature-extraction''': SegformerModel, '''image-classification''': SegformerForImageClassification, '''image-segmentation''': SegformerForSemanticSegmentation, } if is_torch_available() else {} ) _lowercase: Tuple = True _lowercase: Union[str, Any] = False _lowercase: Dict = False _lowercase: Optional[Any] = False def lowercase__ ( self : Tuple ) -> Any: _lowerCAmelCase = SegformerModelTester(self ) _lowerCAmelCase = SegformerConfigTester(self , config_class=__snake_case ) def lowercase__ ( self : Optional[Any] ) -> Dict: self.config_tester.run_common_tests() def lowercase__ ( self : int ) -> Union[str, Any]: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def lowercase__ ( self : Dict ) -> int: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*__snake_case ) def lowercase__ ( self : Dict ) -> Dict: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*__snake_case ) @unittest.skip("""SegFormer does not use inputs_embeds""" ) def lowercase__ ( self : int ) -> Union[str, Any]: pass @unittest.skip("""SegFormer does not have get_input_embeddings method and get_output_embeddings methods""" ) def lowercase__ ( self : Optional[int] ) -> int: pass def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase = model_class(__snake_case ) _lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCAmelCase = [*signature.parameters.keys()] _lowerCAmelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __snake_case ) def lowercase__ ( self : Tuple ) -> Tuple: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase = True for model_class in self.all_model_classes: _lowerCAmelCase = True _lowerCAmelCase = False _lowerCAmelCase = True _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) _lowerCAmelCase = outputs.attentions _lowerCAmelCase = sum(self.model_tester.depths ) self.assertEqual(len(__snake_case ) , __snake_case ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _lowerCAmelCase = True _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) _lowerCAmelCase = outputs.attentions self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first attentions (first block, first layer) _lowerCAmelCase = (self.model_tester.image_size // 4) ** 2 _lowerCAmelCase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) _lowerCAmelCase = (self.model_tester.image_size // 32) ** 2 _lowerCAmelCase = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) _lowerCAmelCase = len(__snake_case ) # Check attention is always last and order is fine _lowerCAmelCase = True _lowerCAmelCase = True _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) self.assertEqual(out_len + 1 , len(__snake_case ) ) _lowerCAmelCase = outputs.attentions self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first attentions (first block, first layer) _lowerCAmelCase = (self.model_tester.image_size // 4) ** 2 _lowerCAmelCase = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def lowercase__ ( self : int ) -> List[str]: def check_hidden_states_output(__snake_case : str , __snake_case : Tuple , __snake_case : Optional[int] ): _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): _lowerCAmelCase = model(**self._prepare_for_class(__snake_case , __snake_case ) ) _lowerCAmelCase = outputs.hidden_states _lowerCAmelCase = self.model_tester.num_encoder_blocks self.assertEqual(len(__snake_case ) , __snake_case ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCAmelCase = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCAmelCase = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) def lowercase__ ( self : Optional[Any] ) -> Any: if not self.model_tester.is_training: return _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCAmelCase = True for model_class in self.all_model_classes: if model_class in get_values(__snake_case ): continue _lowerCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.train() _lowerCAmelCase = self._prepare_for_class(__snake_case , __snake_case , return_labels=__snake_case ) _lowerCAmelCase = model(**__snake_case ).loss loss.backward() @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowercase__ ( self : Tuple ) -> Dict: pass @slow def lowercase__ ( self : str ) -> Optional[int]: for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase = SegformerModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class UpperCAmelCase ( unittest.TestCase ): @slow def lowercase__ ( self : Union[str, Any] ) -> Any: # only resize + normalize _lowerCAmelCase = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) _lowerCAmelCase = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( __snake_case ) _lowerCAmelCase = prepare_img() _lowerCAmelCase = image_processor(images=__snake_case , return_tensors="""pt""" ) _lowerCAmelCase = encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = torch.Size((1, model.config.num_labels, 1_28, 1_28) ) self.assertEqual(outputs.logits.shape , __snake_case ) _lowerCAmelCase = torch.tensor( [ [[-4.63_10, -5.52_32, -6.23_56], [-5.19_21, -6.14_44, -6.59_96], [-5.44_24, -6.27_90, -6.75_74]], [[-12.13_91, -13.31_22, -13.95_54], [-12.87_32, -13.93_52, -14.35_63], [-12.94_38, -13.82_26, -14.25_13]], [[-12.51_34, -13.46_86, -14.49_15], [-12.86_69, -14.43_43, -14.77_58], [-13.25_23, -14.58_19, -15.06_94]], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def lowercase__ ( self : Optional[Any] ) -> Any: # only resize + normalize _lowerCAmelCase = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) _lowerCAmelCase = SegformerForSemanticSegmentation.from_pretrained( """nvidia/segformer-b1-finetuned-cityscapes-1024-1024""" ).to(__snake_case ) _lowerCAmelCase = prepare_img() _lowerCAmelCase = image_processor(images=__snake_case , return_tensors="""pt""" ) _lowerCAmelCase = encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = torch.Size((1, model.config.num_labels, 1_28, 1_28) ) self.assertEqual(outputs.logits.shape , __snake_case ) _lowerCAmelCase = torch.tensor( [ [[-13.57_48, -13.91_11, -12.65_00], [-14.35_00, -15.36_83, -14.23_28], [-14.75_32, -16.04_24, -15.60_87]], [[-17.16_51, -15.87_25, -12.96_53], [-17.25_80, -17.37_18, -14.82_23], [-16.60_58, -16.87_83, -16.74_52]], [[-3.64_56, -3.02_09, -1.42_03], [-3.07_97, -3.19_59, -2.00_00], [-1.87_57, -1.92_17, -1.69_97]], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , __snake_case , atol=1E-1 ) ) @slow def lowercase__ ( self : Any ) -> str: # only resize + normalize _lowerCAmelCase = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=__snake_case , align=__snake_case , do_random_crop=__snake_case ) _lowerCAmelCase = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( __snake_case ) _lowerCAmelCase = prepare_img() _lowerCAmelCase = image_processor(images=__snake_case , return_tensors="""pt""" ) _lowerCAmelCase = encoded_inputs.pixel_values.to(__snake_case ) with torch.no_grad(): _lowerCAmelCase = model(__snake_case ) _lowerCAmelCase = outputs.logits.detach().cpu() _lowerCAmelCase = image_processor.post_process_semantic_segmentation(outputs=__snake_case , target_sizes=[(5_00, 3_00)] ) _lowerCAmelCase = torch.Size((5_00, 3_00) ) self.assertEqual(segmentation[0].shape , __snake_case ) _lowerCAmelCase = image_processor.post_process_semantic_segmentation(outputs=__snake_case ) _lowerCAmelCase = torch.Size((1_28, 1_28) ) self.assertEqual(segmentation[0].shape , __snake_case )
70
0
'''simple docstring''' import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : List[Any] ): '''simple docstring''' _A = logging.get_logger() # the current default level is logging.WARNING _A = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(__snake_case ) def lowerCAmelCase ( self : int ): '''simple docstring''' _A = logging.get_verbosity() _A = logging.get_logger("transformers.models.bart.tokenization_bart" ) _A = "Testing 1, 2, 3" # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(__snake_case ) as cl: logger.warning(__snake_case ) self.assertEqual(cl.out , msg + "\n" ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(__snake_case ) as cl: logger.warning(__snake_case ) self.assertEqual(cl.out , "" ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(__snake_case ) as cl: logger.warning(__snake_case ) self.assertEqual(cl.out , msg + "\n" ) # restore to the original level logging.set_verbosity(__snake_case ) @mockenv(TRANSFORMERS_VERBOSITY="error" ) def lowerCAmelCase ( self : Tuple ): '''simple docstring''' transformers.utils.logging._reset_library_root_logger() # this action activates the env var _A = logging.get_logger("transformers.models.bart.tokenization_bart" ) _A = os.getenv("TRANSFORMERS_VERBOSITY" , __snake_case ) _A = logging.log_levels[env_level_str] _A = logging.get_verbosity() self.assertEqual( __snake_case , __snake_case , f'''TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}''' , ) # restore to the original level _A = "" transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY="super-error" ) def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' transformers.utils.logging._reset_library_root_logger() _A = logging.logging.getLogger() with CaptureLogger(__snake_case ) as cl: # this action activates the env var logging.get_logger("transformers.models.bart.tokenization_bart" ) self.assertIn("Unknown option TRANSFORMERS_VERBOSITY=super-error" , cl.out ) # no need to restore as nothing was changed def lowerCAmelCase ( self : Dict ): '''simple docstring''' transformers.utils.logging._reset_library_root_logger() _A = logging.get_logger("transformers.models.bart.tokenization_bart" ) _A = "Testing 1, 2, 3" with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="1" ): # nothing should be logged as env var disables this method with CaptureLogger(__snake_case ) as cl: logger.warning_advice(__snake_case ) self.assertEqual(cl.out , "" ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="" ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(__snake_case ) as cl: logger.warning_advice(__snake_case ) self.assertEqual(cl.out , msg + "\n" ) def __lowercase ( ) -> Optional[Any]: '''simple docstring''' disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
79
'''simple docstring''' from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class UpperCAmelCase : _lowercase: List[str] _lowercase: Optional[str] = None # Automatically constructed _lowercase: ClassVar[str] = "dict" _lowercase: ClassVar[Any] = None _lowercase: str = field(default='''Translation''' , init=snake_case_ , repr=snake_case_ ) def __call__( self : Optional[int] ) -> Optional[int]: return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def lowercase__ ( self : Union[str, Any] ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value("""string""" ) for k in sorted(self.languages )} @dataclass class UpperCAmelCase : _lowercase: Optional[List] = None _lowercase: Optional[int] = None _lowercase: Optional[str] = None # Automatically constructed _lowercase: ClassVar[str] = "dict" _lowercase: ClassVar[Any] = None _lowercase: str = field(default='''TranslationVariableLanguages''' , init=snake_case_ , repr=snake_case_ ) def lowercase__ ( self : Any ) -> Optional[Any]: _lowerCAmelCase = sorted(set(self.languages ) ) if self.languages else None _lowerCAmelCase = len(self.languages ) if self.languages else None def __call__( self : List[str] ) -> Optional[Any]: return pa.struct({"""language""": pa.list_(pa.string() ), """translation""": pa.list_(pa.string() )} ) def lowercase__ ( self : Optional[Any] , __snake_case : Tuple ) -> Any: _lowerCAmelCase = set(self.languages ) if self.languages and set(__snake_case ) - lang_set: raise ValueError( f"Some languages in example ({', '.join(sorted(set(__snake_case ) - lang_set ) )}) are not in valid set ({', '.join(__snake_case )})." ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. _lowerCAmelCase = [] for lang, text in translation_dict.items(): if isinstance(__snake_case , __snake_case ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. _lowerCAmelCase , _lowerCAmelCase = zip(*sorted(__snake_case ) ) return {"language": languages, "translation": translations} def lowercase__ ( self : str ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value("""string""" ) ), "translation": Sequence(Value("""string""" ) ), }
70
0
import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm A_ : str = re.compile('[^A-Za-z_0-9]') # parameters used in DuplicationIndex A_ : Any = 10 A_ : Optional[int] = 256 def __a ( SCREAMING_SNAKE_CASE ) -> Optional[Any]: '''simple docstring''' if len(SCREAMING_SNAKE_CASE ) < MIN_NUM_TOKENS: return None __UpperCAmelCase = MinHash(num_perm=SCREAMING_SNAKE_CASE ) for token in set(SCREAMING_SNAKE_CASE ): min_hash.update(token.encode() ) return min_hash def __a ( SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' return {t for t in NON_ALPHA.split(SCREAMING_SNAKE_CASE ) if len(t.strip() ) > 0} class A_ : '''simple docstring''' def __init__(self , *, lowercase__ = 0.85 , ) -> List[Any]: __UpperCAmelCase = duplication_jaccard_threshold __UpperCAmelCase = NUM_PERM __UpperCAmelCase = MinHashLSH(threshold=self._duplication_jaccard_threshold , num_perm=self._num_perm ) __UpperCAmelCase = defaultdict(__snake_case ) def lowerCAmelCase_ (self , lowercase__ , lowercase__ ) -> None: __UpperCAmelCase = self._index.query(__snake_case ) if code_key in self._index.keys: print(F'''Duplicate key {code_key}''' ) return self._index.insert(__snake_case , __snake_case ) if len(__snake_case ) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(__snake_case ) break else: self._duplicate_clusters[close_duplicates[0]].add(__snake_case ) def lowerCAmelCase_ (self ) -> List[List[Dict]]: __UpperCAmelCase = [] for base, duplicates in self._duplicate_clusters.items(): __UpperCAmelCase = [base] + list(__snake_case ) # reformat the cluster to be a list of dict __UpperCAmelCase = [{'''base_index''': el[0], '''repo_name''': el[1], '''path''': el[2]} for el in cluster] duplicate_clusters.append(__snake_case ) return duplicate_clusters def lowerCAmelCase_ (self , lowercase__ ) -> None: __UpperCAmelCase = self.get_duplicate_clusters() with open(__snake_case , '''w''' ) as f: json.dump(__snake_case , __snake_case ) def __a ( SCREAMING_SNAKE_CASE ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase , __UpperCAmelCase = element __UpperCAmelCase = get_min_hash([t for t in NON_ALPHA.split(data['''content'''] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def __a ( SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(SCREAMING_SNAKE_CASE , max_queue_size=1_0_0_0_0 ) , chunksize=1_0_0 , ): if data is not None: yield data def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' __UpperCAmelCase = DuplicationIndex(duplication_jaccard_threshold=SCREAMING_SNAKE_CASE ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(SCREAMING_SNAKE_CASE ) ) , max_queue_size=1_0_0 ) ): di.add(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' __UpperCAmelCase = get_tokens(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = get_tokens(SCREAMING_SNAKE_CASE ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) A_ : Tuple = None def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase = [] for elementa in cluster: __UpperCAmelCase = _shared_dataset[elementa['''base_index''']]['''content'''] for elementa in extremes: __UpperCAmelCase = _shared_dataset[elementa['''base_index''']]['''content'''] if jaccard_similarity(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) >= jaccard_threshold: elementa["copies"] += 1 break else: __UpperCAmelCase = 1 extremes.append(SCREAMING_SNAKE_CASE ) return extremes def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Dict: '''simple docstring''' global _shared_dataset __UpperCAmelCase = dataset __UpperCAmelCase = [] __UpperCAmelCase = partial(_find_cluster_extremes_shared , jaccard_threshold=SCREAMING_SNAKE_CASE ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) , total=len(SCREAMING_SNAKE_CASE ) , ): extremes_list.append(SCREAMING_SNAKE_CASE ) return extremes_list def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 0.85 ) -> Tuple: '''simple docstring''' __UpperCAmelCase = make_duplicate_clusters(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __UpperCAmelCase = {x['''base_index'''] for cluster in duplicate_clusters for x in cluster} __UpperCAmelCase = {} __UpperCAmelCase = find_extremes(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for extremes in extremes_clusters: for element in extremes: __UpperCAmelCase = element __UpperCAmelCase = duplicate_indices - set(extreme_dict.keys() ) __UpperCAmelCase = dataset.filter(lambda SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : idx not in remove_indices , with_indices=SCREAMING_SNAKE_CASE ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: __UpperCAmelCase = element['''base_index'''] in extreme_dict if element["is_extreme"]: __UpperCAmelCase = extreme_dict[element['''base_index''']]['''copies'''] print(f'''Original dataset size: {len(SCREAMING_SNAKE_CASE )}''' ) print(f'''Number of duplicate clusters: {len(SCREAMING_SNAKE_CASE )}''' ) print(f'''Files in duplicate cluster: {len(SCREAMING_SNAKE_CASE )}''' ) print(f'''Unique files in duplicate cluster: {len(SCREAMING_SNAKE_CASE )}''' ) print(f'''Filtered dataset size: {len(SCREAMING_SNAKE_CASE )}''' ) return ds_filter, duplicate_clusters
333
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer A__ : List[str] =logging.get_logger(__name__) A__ : Any ={'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A__ : Any ={ '''vocab_file''': { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt''' ), } } A__ : Optional[int] ={ '''junnyu/roformer_chinese_small''': 15_36, '''junnyu/roformer_chinese_base''': 15_36, '''junnyu/roformer_chinese_char_small''': 5_12, '''junnyu/roformer_chinese_char_base''': 5_12, '''junnyu/roformer_small_discriminator''': 1_28, '''junnyu/roformer_small_generator''': 1_28, } A__ : Optional[int] ={ '''junnyu/roformer_chinese_small''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_base''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_char_small''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_char_base''': {'''do_lower_case''': True}, '''junnyu/roformer_small_discriminator''': {'''do_lower_case''': True}, '''junnyu/roformer_small_generator''': {'''do_lower_case''': True}, } class UpperCAmelCase ( snake_case_ ): _lowercase: Optional[Any] = VOCAB_FILES_NAMES _lowercase: Tuple = PRETRAINED_VOCAB_FILES_MAP _lowercase: Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase: str = PRETRAINED_INIT_CONFIGURATION _lowercase: List[Any] = RoFormerTokenizer def __init__( self : Dict , __snake_case : str=None , __snake_case : Tuple=None , __snake_case : List[Any]=True , __snake_case : str="[UNK]" , __snake_case : Tuple="[SEP]" , __snake_case : str="[PAD]" , __snake_case : str="[CLS]" , __snake_case : Any="[MASK]" , __snake_case : Dict=True , __snake_case : str=None , **__snake_case : Optional[Any] , ) -> Union[str, Any]: super().__init__( __snake_case , tokenizer_file=__snake_case , do_lower_case=__snake_case , unk_token=__snake_case , sep_token=__snake_case , pad_token=__snake_case , cls_token=__snake_case , mask_token=__snake_case , tokenize_chinese_chars=__snake_case , strip_accents=__snake_case , **__snake_case , ) _lowerCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get("""lowercase""" , __snake_case ) != do_lower_case or pre_tok_state.get("""strip_accents""" , __snake_case ) != strip_accents ): _lowerCAmelCase = getattr(__snake_case , pre_tok_state.pop("""type""" ) ) _lowerCAmelCase = do_lower_case _lowerCAmelCase = strip_accents _lowerCAmelCase = pre_tok_class(**__snake_case ) _lowerCAmelCase = do_lower_case def __getstate__( self : int ) -> Optional[int]: _lowerCAmelCase = self.__dict__.copy() _lowerCAmelCase = BertPreTokenizer() return state def __setstate__( self : Tuple , __snake_case : Tuple ) -> List[str]: _lowerCAmelCase = d _lowerCAmelCase = self.__dict__["""_tokenizer"""].get_vocab() _lowerCAmelCase = PreTokenizer.custom(JiebaPreTokenizer(__snake_case ) ) def lowercase__ ( self : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[int]=None ) -> Optional[Any]: _lowerCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase__ ( self : List[str] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: _lowerCAmelCase = [self.sep_token_id] _lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase__ ( self : int , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: _lowerCAmelCase = self._tokenizer.model.save(__snake_case , name=__snake_case ) return tuple(__snake_case ) def lowercase__ ( self : Dict , __snake_case : Dict , __snake_case : int=None , __snake_case : List[Any]=None , __snake_case : List[Any]=False , **__snake_case : Dict , ) -> str: _lowerCAmelCase = BertPreTokenizer() return super().save_pretrained(__snake_case , __snake_case , __snake_case , __snake_case , **__snake_case )
70
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class A_ : def __init__( self , _A , _A=2 , _A=True , _A=False , _A=1_0 , _A=3 , _A=3_2 * 4 , _A=3_2 * 6 , _A=4 , _A=3_2 , ): '''simple docstring''' UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = is_training UpperCAmelCase = use_auxiliary_loss UpperCAmelCase = num_queries UpperCAmelCase = num_channels UpperCAmelCase = min_size UpperCAmelCase = max_size UpperCAmelCase = num_labels UpperCAmelCase = mask_feature_size def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( __snake_case ) UpperCAmelCase = torch.ones([self.batch_size, self.min_size, self.max_size] , device=__snake_case ) UpperCAmelCase = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=__snake_case ) > 0.5 ).float() UpperCAmelCase = (torch.rand((self.batch_size, self.num_labels) , device=__snake_case ) > 0.5).long() UpperCAmelCase = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _lowercase ( self ): '''simple docstring''' return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=1_2_8 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def _lowercase ( self , _A , _A ): '''simple docstring''' UpperCAmelCase = output.encoder_hidden_states UpperCAmelCase = output.pixel_decoder_hidden_states UpperCAmelCase = output.transformer_decoder_hidden_states self.parent.assertTrue(len(__snake_case ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__snake_case ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(__snake_case ) , config.decoder_config.decoder_layers ) def _lowercase ( self , _A , _A , _A , _A=False ): '''simple docstring''' with torch.no_grad(): UpperCAmelCase = MaskFormerModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase = model(pixel_values=__snake_case , pixel_mask=__snake_case ) UpperCAmelCase = model(__snake_case , output_hidden_states=__snake_case ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(__snake_case , __snake_case ) def _lowercase ( self , _A , _A , _A , _A , _A ): '''simple docstring''' UpperCAmelCase = MaskFormerForInstanceSegmentation(config=__snake_case ) model.to(__snake_case ) model.eval() def comm_check_on_output(_A ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): UpperCAmelCase = model(pixel_values=__snake_case , pixel_mask=__snake_case ) UpperCAmelCase = model(__snake_case ) comm_check_on_output(__snake_case ) UpperCAmelCase = model( pixel_values=__snake_case , pixel_mask=__snake_case , mask_labels=__snake_case , class_labels=__snake_case ) comm_check_on_output(__snake_case ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class A_ (snake_case_ , snake_case_ , unittest.TestCase ): UpperCAmelCase__ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () UpperCAmelCase__ = ( {'''feature-extraction''': MaskFormerModel, '''image-segmentation''': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False UpperCAmelCase__ = False def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = MaskFormerModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=__snake_case , has_text_modality=__snake_case ) def _lowercase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__snake_case , **__snake_case , output_hidden_states=__snake_case ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*__snake_case ) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip(reason='''MaskFormer is not a generative model''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip(reason='''MaskFormer does not use token embeddings''' ) def _lowercase ( self ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def _lowercase ( self ): '''simple docstring''' pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def _lowercase ( self ): '''simple docstring''' pass def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__snake_case ) UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __snake_case ) @slow def _lowercase ( self ): '''simple docstring''' for model_name in ["facebook/maskformer-swin-small-coco"]: UpperCAmelCase = MaskFormerModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = (self.model_tester.min_size,) * 2 UpperCAmelCase = { '''pixel_values''': torch.randn((2, 3, *size) , device=__snake_case ), '''mask_labels''': torch.randn((2, 1_0, *size) , device=__snake_case ), '''class_labels''': torch.zeros(2 , 1_0 , device=__snake_case ).long(), } UpperCAmelCase = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(__snake_case ) UpperCAmelCase = model(**__snake_case ) self.assertTrue(outputs.loss is not None ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(__snake_case , **__snake_case , output_hidden_states=__snake_case ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(__snake_case ).to(__snake_case ) UpperCAmelCase = model(**__snake_case , output_attentions=__snake_case ) self.assertTrue(outputs.attentions is not None ) def _lowercase ( self ): '''simple docstring''' if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss UpperCAmelCase = self.all_model_classes[1] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs() UpperCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.train() UpperCAmelCase = model(__snake_case , mask_labels=__snake_case , class_labels=__snake_case ).loss loss.backward() def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = self.all_model_classes[1] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs() UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = model_class(__snake_case ) model.to(__snake_case ) model.train() UpperCAmelCase = model(__snake_case , mask_labels=__snake_case , class_labels=__snake_case ) UpperCAmelCase = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() UpperCAmelCase = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't UpperCAmelCase = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() UpperCAmelCase = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=__snake_case ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __A : int = 1E-4 def __SCREAMING_SNAKE_CASE ( ) -> int: '''simple docstring''' UpperCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class A_ (unittest.TestCase ): @cached_property def _lowercase ( self ): '''simple docstring''' return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''' ) if is_vision_available() else None ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''' ).to(__snake_case ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(__snake_case , return_tensors='''pt''' ).to(__snake_case ) UpperCAmelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(__snake_case , (1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) UpperCAmelCase = torch.tensor( [[-0.04_82, 0.92_28, 0.49_51], [-0.25_47, 0.80_17, 0.85_27], [-0.00_69, 0.33_85, -0.00_89]] ).to(__snake_case ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) UpperCAmelCase = torch.tensor( [[-0.84_22, -0.84_34, -0.97_18], [-1.01_44, -0.55_65, -0.41_95], [-1.00_38, -0.44_84, -0.19_61]] ).to(__snake_case ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) UpperCAmelCase = torch.tensor( [[0.28_52, -0.01_59, 0.97_35], [0.62_54, 0.18_58, 0.85_29], [-0.06_80, -0.41_16, 1.84_13]] ).to(__snake_case ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , __snake_case , atol=__snake_case ) ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(__snake_case ) .eval() ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(__snake_case , return_tensors='''pt''' ).to(__snake_case ) UpperCAmelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(__snake_case , (1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) # masks_queries_logits UpperCAmelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) UpperCAmelCase = [ [-1.3_73_71_24, -1.7_72_49_37, -1.9_36_42_33], [-1.5_97_72_81, -1.9_86_79_39, -2.1_52_36_95], [-1.5_79_53_98, -1.9_26_98_32, -2.09_39_42], ] UpperCAmelCase = torch.tensor(__snake_case ).to(__snake_case ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) # class_queries_logits UpperCAmelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase = torch.tensor( [ [1.6512E00, -5.2572E00, -3.3519E00], [3.6169E-02, -5.9025E00, -2.9313E00], [1.0766E-04, -7.7630E00, -5.1263E00], ] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __snake_case , atol=__snake_case ) ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''' ) .to(__snake_case ) .eval() ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(__snake_case , return_tensors='''pt''' ).to(__snake_case ) UpperCAmelCase = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 3_2) == 0 and (inputs_shape[-2] % 3_2) == 0 ) # check size self.assertEqual(__snake_case , (1, 3, 8_0_0, 1_0_8_8) ) with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) # masks_queries_logits UpperCAmelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) UpperCAmelCase = [[-0.90_46, -2.63_66, -4.60_62], [-3.41_79, -5.78_90, -8.80_57], [-4.91_79, -7.65_60, -10.77_11]] UpperCAmelCase = torch.tensor(__snake_case ).to(__snake_case ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , __snake_case , atol=__snake_case ) ) # class_queries_logits UpperCAmelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase = torch.tensor( [[4.71_88, -3.25_85, -2.88_57], [6.68_71, -2.91_81, -1.24_87], [7.24_49, -2.27_64, -2.18_74]] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , __snake_case , atol=__snake_case ) ) def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(__snake_case ) .eval() ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = image_processor( [np.zeros((3, 8_0_0, 1_3_3_3) ), np.zeros((3, 8_0_0, 1_3_3_3) )] , segmentation_maps=[np.zeros((3_8_4, 3_8_4) ).astype(np.floataa ), np.zeros((3_8_4, 3_8_4) ).astype(np.floataa )] , return_tensors='''pt''' , ) UpperCAmelCase = inputs['''pixel_values'''].to(__snake_case ) UpperCAmelCase = [el.to(__snake_case ) for el in inputs['''mask_labels''']] UpperCAmelCase = [el.to(__snake_case ) for el in inputs['''class_labels''']] with torch.no_grad(): UpperCAmelCase = model(**__snake_case ) self.assertTrue(outputs.loss is not None )
273
'''simple docstring''' # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class UpperCAmelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Optional[int] = StableDiffusionControlNetImgaImgPipeline _lowercase: Tuple = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Tuple = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'''control_image'''} ) _lowercase: Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase__ ( self : List[str] ) -> List[str]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Any , __snake_case : str , __snake_case : Any=0 ) -> str: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ) _lowerCAmelCase = floats_tensor(control_image.shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : Optional[int] ) -> List[Any]: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Any = StableDiffusionControlNetImgaImgPipeline _lowercase: Dict = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Any = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(__snake_case : Optional[Any] ): if isinstance(__snake_case , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = MultiControlNetModel([controlneta, controlneta] ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Tuple , __snake_case : int , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), ] _lowerCAmelCase = floats_tensor(control_image[0].shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : List[str] ) -> Dict: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) _lowerCAmelCase = 10.0 _lowerCAmelCase = 4 _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 def lowercase__ ( self : int ) -> str: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Optional[Any] ) -> Dict: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : int ) -> str: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(__snake_case ) except NotImplementedError: pass @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): def lowercase__ ( self : Union[str, Any] ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : List[str] ) -> Any: _lowerCAmelCase = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) _lowerCAmelCase = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=__snake_case , controlnet=__snake_case ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase = """evil space-punk bird""" _lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = pipe( __snake_case , __snake_case , control_image=__snake_case , generator=__snake_case , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) _lowerCAmelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) _lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9E-2
70
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { '''SCUT-DLVCLab/lilt-roberta-en-base''': ( '''https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json''' ), } class lowerCamelCase_ ( snake_case_ ): """simple docstring""" a_ ='''lilt''' def __init__( self : str , _a : Any=3_0522 , _a : str=768 , _a : Any=12 , _a : List[str]=12 , _a : Union[str, Any]=3072 , _a : Tuple="gelu" , _a : List[str]=0.1 , _a : Union[str, Any]=0.1 , _a : Dict=512 , _a : Optional[int]=2 , _a : int=0.02 , _a : Any=1e-12 , _a : Optional[int]=0 , _a : List[str]="absolute" , _a : Dict=None , _a : List[str]=4 , _a : List[Any]=1024 , **_a : List[Any] , ) -> List[str]: super().__init__(pad_token_id=__snake_case , **__snake_case ) __lowerCamelCase : Dict = vocab_size __lowerCamelCase : Union[str, Any] = hidden_size __lowerCamelCase : List[Any] = num_hidden_layers __lowerCamelCase : Tuple = num_attention_heads __lowerCamelCase : List[Any] = hidden_act __lowerCamelCase : Optional[int] = intermediate_size __lowerCamelCase : Union[str, Any] = hidden_dropout_prob __lowerCamelCase : int = attention_probs_dropout_prob __lowerCamelCase : str = max_position_embeddings __lowerCamelCase : Optional[Any] = type_vocab_size __lowerCamelCase : str = initializer_range __lowerCamelCase : Tuple = layer_norm_eps __lowerCamelCase : Any = position_embedding_type __lowerCamelCase : str = classifier_dropout __lowerCamelCase : int = channel_shrink_ratio __lowerCamelCase : Optional[Any] = max_ad_position_embeddings
208
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A__ : List[Any] =logging.get_logger(__name__) A__ : Any =torch.device('''cpu''') def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" _lowerCAmelCase = Image.open(requests.get(lowerCAmelCase , stream=lowerCAmelCase ).raw ) return im def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = dct.pop(lowerCAmelCase ) _lowerCAmelCase = val def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [] for k in state_dict.keys(): _lowerCAmelCase = k if ".pwconv" in k: _lowerCAmelCase = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _lowerCAmelCase = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _lowerCAmelCase = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _lowerCAmelCase = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _lowerCAmelCase = k_new.split(""".""" ) if ls[2].isdigit(): _lowerCAmelCase = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _lowerCAmelCase = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _lowerCAmelCase = 10_00 _lowerCAmelCase = """huggingface/label-files""" _lowerCAmelCase = """imagenet-1k-id2label.json""" _lowerCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) _lowerCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} _lowerCAmelCase = idalabel _lowerCAmelCase = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _lowerCAmelCase = [3, 3, 6, 4] _lowerCAmelCase = [48, 56, 1_12, 2_20] elif swiftformer_name == "swiftformer_s": _lowerCAmelCase = [3, 3, 9, 6] _lowerCAmelCase = [48, 64, 1_68, 2_24] elif swiftformer_name == "swiftformer_l1": _lowerCAmelCase = [4, 3, 10, 5] _lowerCAmelCase = [48, 96, 1_92, 3_84] elif swiftformer_name == "swiftformer_l3": _lowerCAmelCase = [4, 4, 12, 6] _lowerCAmelCase = [64, 1_28, 3_20, 5_12] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _lowerCAmelCase = torch.hub.load_state_dict_from_url(lowerCAmelCase , map_location="""cpu""" , check_hash=lowerCAmelCase ) else: _lowerCAmelCase = torch.load(lowerCAmelCase , map_location="""cpu""" ) _lowerCAmelCase = checkpoint _lowerCAmelCase = create_rename_keys(lowerCAmelCase ) for rename_key_src, rename_key_dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) # load HuggingFace model _lowerCAmelCase = SwiftFormerForImageClassification(lowerCAmelCase ).eval() hf_model.load_state_dict(lowerCAmelCase ) # prepare test inputs _lowerCAmelCase = prepare_img() _lowerCAmelCase = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _lowerCAmelCase = processor(images=lowerCAmelCase , return_tensors="""pt""" ) # compare outputs from both models _lowerCAmelCase = get_expected_output(lowerCAmelCase ) _lowerCAmelCase = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 10_00] ) assert torch.allclose(hf_logits[0, 0:5] , lowerCAmelCase , atol=1e-3 ) Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(f"Saving model {swiftformer_name} to {pytorch_dump_folder_path}" ) hf_model.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": A__ : str =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swiftformer_name''', default='''swiftformer_xs''', choices=['''swiftformer_xs''', '''swiftformer_s''', '''swiftformer_l1''', '''swiftformer_l3'''], type=str, help='''Name of the SwiftFormer model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''./converted_outputs/''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--original_ckpt''', default=None, type=str, help='''Path to the original model checkpoint.''') A__ : Tuple =parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
70
0
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_50, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_00, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class lowercase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase__ ( self ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding='''utf-8''' , check=__snake_case , ) assert hasattr(self , '''env''' ) def lowerCAmelCase__ ( self , UpperCamelCase_=1 ): '''simple docstring''' return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-single''' , instance_count=__snake_case , instance_type=self.instance_type , debugger_hook_config=__snake_case , hyperparameters={**self.env.hyperparameters, '''model_name_or_path''': self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version='''py36''' , ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' TrainingJobAnalytics(__snake_case ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = self.create_estimator() # run training estimator.fit() # result dataframe UpperCamelCase__ :List[Any] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCamelCase__ :Optional[int] = list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value'''] ) UpperCamelCase__ :Tuple = list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value'''] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCamelCase__ :Optional[Any] = ( Session().describe_training_job(estimator.latest_training_job.name ).get('''TrainingTimeInSeconds''' , 999999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy ) assert all(t <= self.results['''eval_loss'''] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , '''w''' ) as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} , __snake_case )
97
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) A__ : List[Any] =pytest.mark.integration @pytest.mark.parametrize("""path""" , ["""paws""", """csv"""] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" inspect_dataset(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = path + """.py""" assert script_name in os.listdir(lowerCAmelCase ) assert "__pycache__" not in os.listdir(lowerCAmelCase ) @pytest.mark.filterwarnings("""ignore:inspect_metric is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.parametrize("""path""" , ["""accuracy"""] ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" inspect_metric(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = path + """.py""" assert script_name in os.listdir(lowerCAmelCase ) assert "__pycache__" not in os.listdir(lowerCAmelCase ) @pytest.mark.parametrize( """path, config_name, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_config_info(lowerCAmelCase , config_name=lowerCAmelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with pytest.raises(lowerCAmelCase ): get_dataset_config_info(lowerCAmelCase , config_name=lowerCAmelCase ) @pytest.mark.parametrize( """path, expected""" , [ ("""squad""", """plain_text"""), ("""acronym_identification""", """default"""), ("""lhoestq/squad""", """plain_text"""), ("""lhoestq/test""", """default"""), ("""lhoestq/demo1""", """lhoestq--demo1"""), ("""dalle-mini/wit""", """dalle-mini--wit"""), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_config_names(lowerCAmelCase ) assert expected in config_names @pytest.mark.parametrize( """path, expected_configs, expected_splits_in_first_config""" , [ ("""squad""", ["""plain_text"""], ["""train""", """validation"""]), ("""dalle-mini/wit""", ["""dalle-mini--wit"""], ["""train"""]), ("""paws""", ["""labeled_final""", """labeled_swap""", """unlabeled_final"""], ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_infos(lowerCAmelCase ) assert list(infos.keys() ) == expected_configs _lowerCAmelCase = expected_configs[0] assert expected_config in infos _lowerCAmelCase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( """path, expected_config, expected_splits""" , [ ("""squad""", """plain_text""", ["""train""", """validation"""]), ("""dalle-mini/wit""", """dalle-mini--wit""", ["""train"""]), ("""paws""", """labeled_final""", ["""train""", """test""", """validation"""]), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_dataset_infos(lowerCAmelCase ) assert expected_config in infos _lowerCAmelCase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( """path, config_name, expected_exception""" , [ ("""paws""", None, ValueError), ] , ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with pytest.raises(lowerCAmelCase ): get_dataset_split_names(lowerCAmelCase , config_name=lowerCAmelCase )
70
0
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCamelCase_ : Dict = logging.get_logger(__name__) @add_end_docstrings(snake_case_ ) class __A ( snake_case_ ): """simple docstring""" def __init__( self , *__A , **__A ) -> Tuple: super().__init__(*__snake_case , **__snake_case ) requires_backends(self , '''vision''' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def SCREAMING_SNAKE_CASE ( self , __A=None ) -> List[Any]: a ={} if top_k is not None: a =top_k return {}, {}, postprocess_params def __call__( self , __A , **__A ) -> int: return super().__call__(__snake_case , **__snake_case ) def SCREAMING_SNAKE_CASE ( self , __A ) -> int: a =load_image(__snake_case ) a =self.image_processor(images=__snake_case , return_tensors=self.framework ) return model_inputs def SCREAMING_SNAKE_CASE ( self , __A ) -> Any: a =self.model(**__snake_case ) return model_outputs def SCREAMING_SNAKE_CASE ( self , __A , __A=5 ) -> Dict: if top_k > self.model.config.num_labels: a =self.model.config.num_labels if self.framework == "pt": a =model_outputs.logits.softmax(-1 )[0] a , a =probs.topk(__snake_case ) elif self.framework == "tf": a =stable_softmax(model_outputs.logits , axis=-1 )[0] a =tf.math.top_k(__snake_case , k=__snake_case ) a , a =topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) a =scores.tolist() a =ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(__snake_case , __snake_case )]
81
'''simple docstring''' from torch import nn def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f"Unsupported activation function: {act_fn}" )
70
0
"""simple docstring""" from __future__ import annotations snake_case_ = list[tuple[int, int]] snake_case_ = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] snake_case_ = ([-1, 0], [0, -1], [1, 0], [0, 1]) # up, left, down, right class A_ : """simple docstring""" def __init__( self :Optional[int] , lowercase_ :int , lowercase_ :int , lowercase_ :int , lowercase_ :int , lowercase_ :float , lowercase_ :Node | None , ) -> Dict: UpperCAmelCase = pos_x UpperCAmelCase = pos_y UpperCAmelCase = (pos_y, pos_x) UpperCAmelCase = goal_x UpperCAmelCase = goal_y UpperCAmelCase = g_cost UpperCAmelCase = parent UpperCAmelCase = self.calculate_heuristic() def UpperCAmelCase__ ( self :str ) -> float: UpperCAmelCase = abs(self.pos_x - self.goal_x ) UpperCAmelCase = abs(self.pos_y - self.goal_y ) return dx + dy def __lt__( self :Optional[int] , lowercase_ :Any ) -> bool: return self.f_cost < other.f_cost class A_ : """simple docstring""" def __init__( self :Optional[int] , lowercase_ :tuple[int, int] , lowercase_ :tuple[int, int] ) -> str: UpperCAmelCase = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , __snake_case ) UpperCAmelCase = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_99_99 , __snake_case ) UpperCAmelCase = [self.start] UpperCAmelCase = [] UpperCAmelCase = False def UpperCAmelCase__ ( self :Optional[Any] ) -> Path | None: while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: UpperCAmelCase = True return self.retrace_path(__snake_case ) self.closed_nodes.append(__snake_case ) UpperCAmelCase = self.get_successors(__snake_case ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(__snake_case ) else: # retrieve the best current path UpperCAmelCase = self.open_nodes.pop(self.open_nodes.index(__snake_case ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(__snake_case ) else: self.open_nodes.append(__snake_case ) if not self.reached: return [self.start.pos] return None def UpperCAmelCase__ ( self :Optional[int] , lowercase_ :Node ) -> list[Node]: UpperCAmelCase = [] for action in delta: UpperCAmelCase = parent.pos_x + action[1] UpperCAmelCase = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(__snake_case ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( __snake_case , __snake_case , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , __snake_case , ) ) return successors def UpperCAmelCase__ ( self :int , lowercase_ :Node | None ) -> Path: UpperCAmelCase = node UpperCAmelCase = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase = current_node.parent path.reverse() return path if __name__ == "__main__": snake_case_ = (0, 0) snake_case_ = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) print("""------""") snake_case_ = GreedyBestFirst(init, goal) snake_case_ = greedy_bf.search() if path: for pos_x, pos_y in path: snake_case_ = 2 for elem in grid: print(elem)
78
'''simple docstring''' # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position A__ : Dict ='''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip A__ : Tuple =concatenate_datasets A__ : Dict =DownloadConfig A__ : int =DownloadManager A__ : Union[str, Any] =DownloadMode A__ : Tuple =DownloadConfig A__ : Optional[Any] =DownloadMode A__ : str =DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
70
0
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass A_ :Tuple = (3, 9, -11, 0, 7, 5, 1, -1) A_ :Optional[int] = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class __A : """simple docstring""" UpperCamelCase__ : int UpperCamelCase__ : Node | None class __A : """simple docstring""" def __init__( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Node | None =None for i in sorted(lowerCamelCase__ , reverse=lowerCamelCase__ ): __UpperCamelCase : Dict =Node(lowerCamelCase__ , self.head ) def __iter__( self ): """simple docstring""" __UpperCamelCase : List[Any] =self.head while node: yield node.data __UpperCamelCase : Tuple =node.next_node def __len__( self ): """simple docstring""" return sum(1 for _ in self ) def __str__( self ): """simple docstring""" return " -> ".join([str(lowerCamelCase__ ) for node in self] ) def A ( a_ ,a_ ) -> SortedLinkedList: return SortedLinkedList(list(a_ ) + list(a_ ) ) if __name__ == "__main__": import doctest doctest.testmod() A_ :Optional[int] = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
71
import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() A_ :List[str] = [ '''word_embeddings_layernorm.weight''', '''word_embeddings_layernorm.bias''', '''input_layernorm.weight''', '''input_layernorm.bias''', '''post_attention_layernorm.weight''', '''post_attention_layernorm.bias''', '''self_attention.dense.bias''', '''mlp.dense_4h_to_h.bias''', '''ln_f.weight''', '''ln_f.bias''', ] A_ :Optional[Any] = [ '''mlp.dense_4h_to_h.weight''', '''self_attention.dense.weight''', ] def A ( a_ ,a_ ) -> str: __UpperCamelCase : Any ={ 'word_embeddings.weight': 'word_embeddings.weight', 'word_embeddings.norm.weight': 'word_embeddings_layernorm.weight', 'word_embeddings.norm.bias': 'word_embeddings_layernorm.bias', 'weight': 'ln_f.weight', 'bias': 'ln_f.bias', } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks __UpperCamelCase : Tuple =int(re.match(r'.*layer_(\d*).*' ,a_ )[1] ) layer_number -= 3 return F'h.{layer_number}.' + key def A ( a_ ) -> Any: if dtype == torch.bool: return 1 / 8 __UpperCamelCase : Dict =re.search(r'[^\d](\d+)$' ,str(a_ ) ) if bit_search is None: raise ValueError(F'`dtype` is not a valid dtype: {dtype}.' ) __UpperCamelCase : Tuple =int(bit_search.groups()[0] ) return bit_size // 8 def A ( a_ ,a_ ,a_ ,a_ ,a_ ) -> Dict: # Construct model if bloom_config_file == "": __UpperCamelCase : List[Any] =BloomConfig() else: __UpperCamelCase : List[str] =BloomConfig.from_json_file(a_ ) if shard_model: __UpperCamelCase : int =os.listdir(a_ ) __UpperCamelCase : Union[str, Any] =sorted(filter(lambda a_ : s.startswith('layer' ) and "model_00" in s ,a_ ) ) __UpperCamelCase : Optional[Any] ={'weight_map': {}, 'metadata': {}} __UpperCamelCase : Dict =0 __UpperCamelCase : int =None __UpperCamelCase : Any =BloomConfig() for j, file in enumerate(a_ ): print('Processing file: {}'.format(a_ ) ) __UpperCamelCase : Optional[int] =None for i in range(a_ ): # load all TP files __UpperCamelCase : Dict =file.replace('model_00' ,F'model_0{i}' ) __UpperCamelCase : Optional[Any] =torch.load(os.path.join(a_ ,a_ ) ,map_location='cpu' ) # Rename keys in the transformers names __UpperCamelCase : int =list(temp.keys() ) for key in keys: __UpperCamelCase : Dict =temp.pop(a_ ) if tensors is None: __UpperCamelCase : Any =temp else: for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel __UpperCamelCase : List[Any] =1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks __UpperCamelCase : Any =torch.cat([tensors[key], temp[key]] ,dim=a_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): __UpperCamelCase : Optional[Any] =tensors[key] / pretraining_tp torch.save( a_ ,os.path.join( a_ ,'pytorch_model_{}-of-{}.bin'.format(str(j + 1 ).zfill(5 ) ,str(len(a_ ) ).zfill(5 ) ) ,) ,) for key in tensors.keys(): __UpperCamelCase : Union[str, Any] =tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: __UpperCamelCase : int ='pytorch_model_{}-of-{}.bin'.format( str(j + 1 ).zfill(5 ) ,str(len(a_ ) ).zfill(5 ) ) __UpperCamelCase : Union[str, Any] =BloomConfig() __UpperCamelCase : Tuple =pytorch_dump_folder_path + '/' + CONFIG_NAME __UpperCamelCase : Optional[int] =total_size with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) with open(os.path.join(a_ ,WEIGHTS_NAME + '.index.json' ) ,'w' ,encoding='utf-8' ) as f: __UpperCamelCase : List[Any] =json.dumps(a_ ,indent=2 ,sort_keys=a_ ) + '\n' f.write(a_ ) else: __UpperCamelCase : List[Any] =BloomModel(a_ ) __UpperCamelCase : Optional[Any] =os.listdir(a_ ) __UpperCamelCase : Dict =sorted(filter(lambda a_ : s.startswith('layer' ) and "model_00" in s ,a_ ) ) __UpperCamelCase : Any =None for i, file in enumerate(a_ ): __UpperCamelCase : Union[str, Any] =None for i in range(a_ ): # load all TP files __UpperCamelCase : Optional[Any] =file.replace('model_00' ,F'model_0{i}' ) __UpperCamelCase : str =torch.load(os.path.join(a_ ,a_ ) ,map_location='cpu' ) # Rename keys in the transformers names __UpperCamelCase : List[str] =list(temp.keys() ) for key in keys: __UpperCamelCase : Union[str, Any] =temp.pop(a_ ) if tensors is None: __UpperCamelCase : Optional[Any] =temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel __UpperCamelCase : Optional[int] =1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks __UpperCamelCase : int =torch.cat([tensors[key], temp[key]] ,dim=a_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): __UpperCamelCase : Dict =tensors[key] / pretraining_tp __UpperCamelCase : str =model.load_state_dict(a_ ,strict=a_ ) assert not other_keys.unexpected_keys, F'The keys {other_keys.unexpected_keys} are unexpected' if missing_keys is None: __UpperCamelCase : str =set(other_keys.missing_keys ) else: __UpperCamelCase : int =missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F'The keys {missing_keys} are missing' # Save pytorch-model os.makedirs(a_ ,exist_ok=a_ ) __UpperCamelCase : Optional[int] =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase : Dict =pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}' ) if config.torch_dtype is not None: __UpperCamelCase : List[str] =model.to(config.torch_dtype ) torch.save(model.state_dict() ,a_ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A_ :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bloom_checkpoint_path''', default=None, type=str, required=True, help='''Path to the Megatron-LM checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--bloom_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--shard_model''', action='''store_true''', help='''An optional setting to shard the output model \nThis enables sharding the converted checkpoint''', ) parser.add_argument( '''--pretraining_tp''', default=4, type=int, help='''Pretraining TP rank that has been used when training the model in Megatron-LM \n''', ) A_ :str = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
71
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL A_ :Any = logging.get_logger(__name__) class __A ( a ): """simple docstring""" UpperCamelCase__ : int =["""pixel_values"""] def __init__( self , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = PIL.Image.BICUBIC , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = 1 / 255 , lowerCamelCase__ = True , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =size if size is not None else {'height': 256, 'width': 256} __UpperCamelCase : List[str] =get_size_dict(lowerCamelCase__ ) __UpperCamelCase : Tuple =crop_size if crop_size is not None else {'height': 224, 'width': 224} __UpperCamelCase : Optional[int] =get_size_dict(lowerCamelCase__ , param_name='crop_size' ) __UpperCamelCase : Tuple =do_resize __UpperCamelCase : Any =size __UpperCamelCase : Dict =resample __UpperCamelCase : Tuple =do_center_crop __UpperCamelCase : int =crop_size __UpperCamelCase : Optional[int] =do_rescale __UpperCamelCase : List[Any] =rescale_factor __UpperCamelCase : List[Any] =do_normalize __UpperCamelCase : Union[str, Any] =image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __UpperCamelCase : Optional[Any] =image_std if image_std is not None else IMAGENET_STANDARD_STD def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = PIL.Image.BICUBIC , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : int =get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return resize( lowerCamelCase__ , size=(size['height'], size['width']) , resample=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[int] =get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(lowerCamelCase__ , size=(size['height'], size['width']) , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" return rescale(lowerCamelCase__ , scale=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" return normalize(lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__=None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = ChannelDimension.FIRST , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[Any] =do_resize if do_resize is not None else self.do_resize __UpperCamelCase : Union[str, Any] =resample if resample is not None else self.resample __UpperCamelCase : Optional[Any] =do_center_crop if do_center_crop is not None else self.do_center_crop __UpperCamelCase : Any =do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase : Dict =rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCamelCase : Optional[Any] =do_normalize if do_normalize is not None else self.do_normalize __UpperCamelCase : Optional[Any] =image_mean if image_mean is not None else self.image_mean __UpperCamelCase : Any =image_std if image_std is not None else self.image_std __UpperCamelCase : Union[str, Any] =size if size is not None else self.size __UpperCamelCase : List[str] =get_size_dict(lowerCamelCase__ ) __UpperCamelCase : Tuple =crop_size if crop_size is not None else self.crop_size __UpperCamelCase : Tuple =get_size_dict(lowerCamelCase__ , param_name='crop_size' ) __UpperCamelCase : List[Any] =make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. __UpperCamelCase : Optional[int] =[to_numpy_array(lowerCamelCase__ ) for image in images] if do_resize: __UpperCamelCase : Optional[Any] =[self.resize(image=lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ ) for image in images] if do_center_crop: __UpperCamelCase : int =[self.center_crop(image=lowerCamelCase__ , size=lowerCamelCase__ ) for image in images] if do_rescale: __UpperCamelCase : int =[self.rescale(image=lowerCamelCase__ , scale=lowerCamelCase__ ) for image in images] if do_normalize: __UpperCamelCase : Optional[int] =[self.normalize(image=lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ ) for image in images] __UpperCamelCase : Optional[Any] =[to_channel_dimension_format(lowerCamelCase__ , lowerCamelCase__ ) for image in images] __UpperCamelCase : List[str] ={'pixel_values': images} return BatchFeature(data=lowerCamelCase__ , tensor_type=lowerCamelCase__ )
71
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class __A : """simple docstring""" UpperCamelCase__ : int =XGLMConfig UpperCamelCase__ : Optional[Any] ={} UpperCamelCase__ : List[str] ="""gelu""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=14 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=99 , lowerCamelCase__=32 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=37 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=512 , lowerCamelCase__=0.02 , ): """simple docstring""" __UpperCamelCase : Tuple =parent __UpperCamelCase : List[str] =batch_size __UpperCamelCase : str =seq_length __UpperCamelCase : Dict =is_training __UpperCamelCase : Tuple =use_input_mask __UpperCamelCase : List[Any] =use_labels __UpperCamelCase : Any =vocab_size __UpperCamelCase : List[Any] =d_model __UpperCamelCase : Optional[int] =num_hidden_layers __UpperCamelCase : List[str] =num_attention_heads __UpperCamelCase : Optional[int] =ffn_dim __UpperCamelCase : str =activation_function __UpperCamelCase : Any =activation_dropout __UpperCamelCase : Optional[int] =attention_dropout __UpperCamelCase : Optional[int] =max_position_embeddings __UpperCamelCase : Any =initializer_range __UpperCamelCase : Dict =None __UpperCamelCase : Optional[int] =0 __UpperCamelCase : Optional[Any] =2 __UpperCamelCase : str =1 def __lowercase ( self ): """simple docstring""" return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) __UpperCamelCase : Union[str, Any] =None if self.use_input_mask: __UpperCamelCase : Dict =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase : Any =self.get_config() __UpperCamelCase : Optional[Any] =floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def __lowercase ( self ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=lowerCamelCase__ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=lowerCamelCase__ , ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) : int =config_and_inputs __UpperCamelCase : Optional[Any] ={ 'input_ids': input_ids, 'head_mask': head_mask, } return config, inputs_dict @require_tf class __A ( a , a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Union[str, Any] =(TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () UpperCamelCase__ : str =(TFXGLMForCausalLM,) if is_tf_available() else () UpperCamelCase__ : Optional[Any] =( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) UpperCamelCase__ : Tuple =False UpperCamelCase__ : Tuple =False UpperCamelCase__ : Optional[Any] =False def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =TFXGLMModelTester(self ) __UpperCamelCase : Dict =ConfigTester(self , config_class=lowerCamelCase__ , n_embd=37 ) def __lowercase ( self ): """simple docstring""" self.config_tester.run_common_tests() @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Optional[Any] =TFXGLMModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def __lowercase ( self ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self , lowerCamelCase__=True ): """simple docstring""" __UpperCamelCase : int =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : List[str] =tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off __UpperCamelCase : str =[2, 268, 9865, 67, 11, 1988, 57252, 9865, 5, 984, 67, 1988, 213838, 1658, 53, 70446, 33, 6657, 278, 1581] # fmt: on __UpperCamelCase : Optional[Any] =model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Union[str, Any] =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) __UpperCamelCase : str =tokenizer('Today is a nice day and' , return_tensors='tf' ) __UpperCamelCase : Union[str, Any] =tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): __UpperCamelCase : Any =model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ , seed=[7, 0] ) __UpperCamelCase : Tuple =tokenizer.decode(output_ids[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : List[Any] =( 'Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due' ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Optional[Any] =XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Optional[Any] ='left' # use different length sentences to test batching __UpperCamelCase : Optional[int] =[ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When', 'Hello, my dog is a little', ] __UpperCamelCase : List[Any] =tokenizer(lowerCamelCase__ , return_tensors='tf' , padding=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =inputs['input_ids'] __UpperCamelCase : Dict =model.generate(input_ids=lowerCamelCase__ , attention_mask=inputs['attention_mask'] , max_new_tokens=12 ) __UpperCamelCase : List[Any] =tokenizer(sentences[0] , return_tensors='tf' ).input_ids __UpperCamelCase : Dict =model.generate(input_ids=lowerCamelCase__ , max_new_tokens=12 ) __UpperCamelCase : Any =tokenizer(sentences[1] , return_tensors='tf' ).input_ids __UpperCamelCase : Optional[Any] =model.generate(input_ids=lowerCamelCase__ , max_new_tokens=12 ) __UpperCamelCase : Optional[int] =tokenizer.batch_decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : int =tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : Any =[ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ' 'a single', 'Hello, my dog is a little bit of a shy one, but he is very friendly', ] self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , [non_padded_sentence, padded_sentence] )
71
1
def A ( a_ ,a_ ,a_ ) -> Tuple: if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(a_ ,n - 1 ,a_ ) * a) % mod else: __UpperCamelCase : Dict =binary_exponentiation(a_ ,n / 2 ,a_ ) return (b * b) % mod # a prime number A_ :str = 701 A_ :Optional[Any] = 1000000000 A_ :int = 10 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
71
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A ( a_ ,a_ ) -> Optional[Any]: # Load checkpoint __UpperCamelCase : int =torch.load(a_ ,map_location='cpu' ) __UpperCamelCase : List[Any] =chkpt['model'] # We have the base model one level deeper than the original XLM repository __UpperCamelCase : str ={} for k, v in state_dict.items(): if "pred_layer" in k: __UpperCamelCase : Optional[Any] =v else: __UpperCamelCase : Optional[Any] =v __UpperCamelCase : List[Any] =chkpt['params'] __UpperCamelCase : str ={n: v for n, v in config.items() if not isinstance(a_ ,(torch.FloatTensor, numpy.ndarray) )} __UpperCamelCase : str =chkpt['dico_word2id'] __UpperCamelCase : Dict ={s + '</w>' if s.find('@@' ) == -1 and i > 13 else s.replace('@@' ,'' ): i for s, i in vocab.items()} # Save pytorch-model __UpperCamelCase : List[Any] =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase : Tuple =pytorch_dump_folder_path + '/' + CONFIG_NAME __UpperCamelCase : Any =pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['vocab_file'] print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(a_ ,a_ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(a_ ,indent=2 ) + '\n' ) print(F'Save vocab file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(a_ ,indent=2 ) + '\n' ) if __name__ == "__main__": A_ :str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xlm_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) A_ :List[Any] = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
71
1
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: A_ :int = None A_ :Optional[int] = logging.get_logger(__name__) A_ :Tuple = {'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} A_ :int = { '''vocab_file''': { '''facebook/mbart-large-en-ro''': ( '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model''' ), '''facebook/mbart-large-cc25''': ( '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/mbart-large-en-ro''': '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json''', '''facebook/mbart-large-cc25''': '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json''', }, } A_ :Union[str, Any] = { '''facebook/mbart-large-en-ro''': 1024, '''facebook/mbart-large-cc25''': 1024, } # fmt: off A_ :int = ['''ar_AR''', '''cs_CZ''', '''de_DE''', '''en_XX''', '''es_XX''', '''et_EE''', '''fi_FI''', '''fr_XX''', '''gu_IN''', '''hi_IN''', '''it_IT''', '''ja_XX''', '''kk_KZ''', '''ko_KR''', '''lt_LT''', '''lv_LV''', '''my_MM''', '''ne_NP''', '''nl_XX''', '''ro_RO''', '''ru_RU''', '''si_LK''', '''tr_TR''', '''vi_VN''', '''zh_CN'''] class __A ( a ): """simple docstring""" UpperCamelCase__ : List[str] =VOCAB_FILES_NAMES UpperCamelCase__ : Union[str, Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Tuple =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : str =["""input_ids""", """attention_mask"""] UpperCamelCase__ : Any =MBartTokenizer UpperCamelCase__ : List[int] =[] UpperCamelCase__ : List[int] =[] def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<mask>" , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Union[str, Any] =AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token super().__init__( vocab_file=lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , src_lang=lowerCamelCase__ , tgt_lang=lowerCamelCase__ , additional_special_tokens=lowerCamelCase__ , **lowerCamelCase__ , ) __UpperCamelCase : int =vocab_file __UpperCamelCase : Tuple =False if not self.vocab_file else True __UpperCamelCase : Dict =FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) __UpperCamelCase : List[str] ={ lang_code: self.convert_tokens_to_ids(lowerCamelCase__ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } __UpperCamelCase : Optional[Any] =src_lang if src_lang is not None else 'en_XX' __UpperCamelCase : str =self.convert_tokens_to_ids(self._src_lang ) __UpperCamelCase : str =tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __lowercase ( self ): """simple docstring""" return self._src_lang @src_lang.setter def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Union[str, Any] =new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : List[Any] =[self.sep_token_id] __UpperCamelCase : Tuple =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) __UpperCamelCase : List[str] =src_lang __UpperCamelCase : List[str] =self(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =self.convert_tokens_to_ids(lowerCamelCase__ ) __UpperCamelCase : Optional[int] =tgt_lang_id return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = "en_XX" , lowerCamelCase__ = None , lowerCamelCase__ = "ro_RO" , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Dict =src_lang __UpperCamelCase : Optional[Any] =tgt_lang return super().prepare_seqaseq_batch(lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang ) def __lowercase ( self ): """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] =self.convert_tokens_to_ids(lowerCamelCase__ ) __UpperCamelCase : Tuple =[] __UpperCamelCase : Optional[Any] =[self.eos_token_id, self.cur_lang_code] __UpperCamelCase : Optional[Any] =self.convert_ids_to_tokens(self.prefix_tokens ) __UpperCamelCase : List[Any] =self.convert_ids_to_tokens(self.suffix_tokens ) __UpperCamelCase : Union[str, Any] =processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.convert_tokens_to_ids(lowerCamelCase__ ) __UpperCamelCase : int =[] __UpperCamelCase : List[str] =[self.eos_token_id, self.cur_lang_code] __UpperCamelCase : Optional[int] =self.convert_ids_to_tokens(self.prefix_tokens ) __UpperCamelCase : Dict =self.convert_ids_to_tokens(self.suffix_tokens ) __UpperCamelCase : List[str] =processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory.' ) return __UpperCamelCase : List[Any] =os.path.join( lowerCamelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file , lowerCamelCase__ ) return (out_vocab_file,)
71
import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __A ( a ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =tempfile.mkdtemp() __UpperCamelCase : Optional[int] =8 # DPR tok __UpperCamelCase : str =[ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase : Optional[Any] =os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase : Optional[int] =[ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase : str =dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) __UpperCamelCase : Optional[int] =['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase : Any ={'unk_token': '<unk>'} __UpperCamelCase : Any =os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Any =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(lowerCamelCase__ ) ) def __lowercase ( self ): """simple docstring""" return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) @require_tokenizers def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =os.path.join(self.tmpdirname , 'rag_tokenizer' ) __UpperCamelCase : Dict =RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) __UpperCamelCase : List[Any] =RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(lowerCamelCase__ ) rag_tokenizer.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : int =RagTokenizer.from_pretrained(lowerCamelCase__ , config=lowerCamelCase__ ) self.assertIsInstance(new_rag_tokenizer.question_encoder , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =RagTokenizer.from_pretrained('facebook/rag-token-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : int =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =RagTokenizer.from_pretrained('facebook/rag-sequence-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : Any =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
71
1
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print('''Googling.....''') A_ :Tuple = '''https://www.google.com/search?q=''' + ''' '''.join(sys.argv[1:]) A_ :str = requests.get(url, headers={'''UserAgent''': UserAgent().random}) # res.raise_for_status() with open('''project1a.html''', '''wb''') as out_file: # only for knowing the class for data in res.iter_content(10000): out_file.write(data) A_ :List[Any] = BeautifulSoup(res.text, '''html.parser''') A_ :List[Any] = list(soup.select('''.eZt8xd'''))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get('''href''')) else: webbrowser.open(f"https://google.com{link.get('href')}")
71
A_ :Optional[int] = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' A_ :Union[str, Any] = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] A_ :Optional[Any] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
71
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A_ :Dict = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Optional[Any] = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :List[str] = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys A_ :List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
71
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration A_ :Optional[Any] = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def A ( a_ ) -> List[Any]: __UpperCamelCase : Any =['layers', 'blocks'] for k in ignore_keys: state_dict.pop(a_ ,a_ ) A_ :int = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def A ( a_ ) -> Union[str, Any]: __UpperCamelCase : str =list(s_dict.keys() ) for key in keys: __UpperCamelCase : str =key for k, v in WHISPER_MAPPING.items(): if k in key: __UpperCamelCase : Optional[Any] =new_key.replace(a_ ,a_ ) print(F'{key} -> {new_key}' ) __UpperCamelCase : Dict =s_dict.pop(a_ ) return s_dict def A ( a_ ) -> Optional[Any]: __UpperCamelCase , __UpperCamelCase : Tuple =emb.weight.shape __UpperCamelCase : Tuple =nn.Linear(a_ ,a_ ,bias=a_ ) __UpperCamelCase : List[Any] =emb.weight.data return lin_layer def A ( a_ ,a_ ) -> bytes: os.makedirs(a_ ,exist_ok=a_ ) __UpperCamelCase : Optional[int] =os.path.basename(a_ ) __UpperCamelCase : Union[str, Any] =url.split('/' )[-2] __UpperCamelCase : Union[str, Any] =os.path.join(a_ ,a_ ) if os.path.exists(a_ ) and not os.path.isfile(a_ ): raise RuntimeError(F'{download_target} exists and is not a regular file' ) if os.path.isfile(a_ ): __UpperCamelCase : str =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(F'{download_target} exists, but the SHA256 checksum does not match; re-downloading the file' ) with urllib.request.urlopen(a_ ) as source, open(a_ ,'wb' ) as output: with tqdm( total=int(source.info().get('Content-Length' ) ) ,ncols=80 ,unit='iB' ,unit_scale=a_ ,unit_divisor=1_024 ) as loop: while True: __UpperCamelCase : Optional[Any] =source.read(8_192 ) if not buffer: break output.write(a_ ) loop.update(len(a_ ) ) __UpperCamelCase : List[Any] =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() != expected_shaaaa: raise RuntimeError( 'Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.' ) return model_bytes def A ( a_ ,a_ ) -> Optional[Any]: if ".pt" not in checkpoint_path: __UpperCamelCase : int =_download(_MODELS[checkpoint_path] ) else: __UpperCamelCase : List[str] =torch.load(a_ ,map_location='cpu' ) __UpperCamelCase : Union[str, Any] =original_checkpoint['dims'] __UpperCamelCase : List[Any] =original_checkpoint['model_state_dict'] __UpperCamelCase : Dict =state_dict['decoder.token_embedding.weight'] remove_ignore_keys_(a_ ) rename_keys(a_ ) __UpperCamelCase : List[str] =True __UpperCamelCase : str =state_dict['decoder.layers.0.fc1.weight'].shape[0] __UpperCamelCase : Optional[int] =WhisperConfig( vocab_size=dimensions['n_vocab'] ,encoder_ffn_dim=a_ ,decoder_ffn_dim=a_ ,num_mel_bins=dimensions['n_mels'] ,d_model=dimensions['n_audio_state'] ,max_target_positions=dimensions['n_text_ctx'] ,encoder_layers=dimensions['n_audio_layer'] ,encoder_attention_heads=dimensions['n_audio_head'] ,decoder_layers=dimensions['n_text_layer'] ,decoder_attention_heads=dimensions['n_text_state'] ,max_source_positions=dimensions['n_audio_ctx'] ,) __UpperCamelCase : List[str] =WhisperForConditionalGeneration(a_ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =model.model.load_state_dict(a_ ,strict=a_ ) if len(a_ ) > 0 and not set(a_ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( 'Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,' F' but all the following weights are missing {missing}' ) if tie_embeds: __UpperCamelCase : Optional[int] =make_linear_from_emb(model.model.decoder.embed_tokens ) else: __UpperCamelCase : List[str] =proj_out_weights model.save_pretrained(a_ ) if __name__ == "__main__": A_ :List[Any] = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') A_ :List[Any] = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
71
1
import os from distutils.util import strtobool def A ( a_ ,a_ ) -> Dict: for e in env_keys: __UpperCamelCase : Union[str, Any] =int(os.environ.get(a_ ,-1 ) ) if val >= 0: return val return default def A ( a_ ,a_=False ) -> Tuple: __UpperCamelCase : Tuple =os.environ.get(a_ ,str(a_ ) ) return strtobool(a_ ) == 1 # As its name indicates `strtobool` actually returns an int... def A ( a_ ,a_="no" ) -> int: __UpperCamelCase : Any =os.environ.get(a_ ,str(a_ ) ) return value
71
import os from datetime import datetime as dt from github import Github A_ :str = [ '''good first issue''', '''feature request''', '''wip''', ] def A ( ) -> Any: __UpperCamelCase : Any =Github(os.environ['GITHUB_TOKEN'] ) __UpperCamelCase : Union[str, Any] =g.get_repo('huggingface/accelerate' ) __UpperCamelCase : Tuple =repo.get_issues(state='open' ) for issue in open_issues: __UpperCamelCase : List[Any] =sorted([comment for comment in issue.get_comments()] ,key=lambda a_ : i.created_at ,reverse=a_ ) __UpperCamelCase : str =comments[0] if len(a_ ) > 0 else None __UpperCamelCase : Any =dt.utcnow() __UpperCamelCase : List[str] =(current_time - issue.updated_at).days __UpperCamelCase : Union[str, Any] =(current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state='closed' ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
71
1
import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __A ( a ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =tempfile.mkdtemp() __UpperCamelCase : Optional[int] =8 # DPR tok __UpperCamelCase : str =[ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase : Optional[Any] =os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase : Optional[int] =[ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase : str =dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) __UpperCamelCase : Optional[int] =['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase : Any ={'unk_token': '<unk>'} __UpperCamelCase : Any =os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Any =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(lowerCamelCase__ ) ) def __lowercase ( self ): """simple docstring""" return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) @require_tokenizers def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =os.path.join(self.tmpdirname , 'rag_tokenizer' ) __UpperCamelCase : Dict =RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) __UpperCamelCase : List[Any] =RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(lowerCamelCase__ ) rag_tokenizer.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : int =RagTokenizer.from_pretrained(lowerCamelCase__ , config=lowerCamelCase__ ) self.assertIsInstance(new_rag_tokenizer.question_encoder , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =RagTokenizer.from_pretrained('facebook/rag-token-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : int =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =RagTokenizer.from_pretrained('facebook/rag-sequence-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : Any =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
71
import re def A ( a_ ) -> bool: __UpperCamelCase : Any =re.compile( r'^(?:0|94|\+94|0{2}94)' r'7(0|1|2|4|5|6|7|8)' r'(-| |)' r'\d{7}$' ) return bool(re.search(a_ ,a_ ) ) if __name__ == "__main__": A_ :List[str] = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
71
1
from ...configuration_utils import PretrainedConfig from ...utils import logging A_ :Any = logging.get_logger(__name__) A_ :int = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[int] ="""vit_msn""" def __init__( self , lowerCamelCase__=768 , lowerCamelCase__=12 , lowerCamelCase__=12 , lowerCamelCase__=3072 , lowerCamelCase__="gelu" , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.02 , lowerCamelCase__=1E-06 , lowerCamelCase__=224 , lowerCamelCase__=16 , lowerCamelCase__=3 , lowerCamelCase__=True , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : int =hidden_size __UpperCamelCase : List[Any] =num_hidden_layers __UpperCamelCase : Union[str, Any] =num_attention_heads __UpperCamelCase : List[str] =intermediate_size __UpperCamelCase : Union[str, Any] =hidden_act __UpperCamelCase : str =hidden_dropout_prob __UpperCamelCase : Union[str, Any] =attention_probs_dropout_prob __UpperCamelCase : Union[str, Any] =initializer_range __UpperCamelCase : Tuple =layer_norm_eps __UpperCamelCase : Optional[Any] =image_size __UpperCamelCase : Optional[int] =patch_size __UpperCamelCase : Any =num_channels __UpperCamelCase : str =qkv_bias
71
A_ :str = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
71
1
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy A_ :int = logging.get_logger(__name__) class __A ( a ): """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] =feature_size __UpperCamelCase : Tuple =sampling_rate __UpperCamelCase : List[Any] =padding_value __UpperCamelCase : List[Any] =kwargs.pop('padding_side' , 'right' ) __UpperCamelCase : Optional[Any] =kwargs.pop('return_attention_mask' , lowerCamelCase__ ) super().__init__(**lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = False , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , ): """simple docstring""" if isinstance(lowerCamelCase__ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): __UpperCamelCase : List[Any] ={ key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( 'You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`' f' to this method that includes {self.model_input_names[0]}, but you provided' f' {list(processed_features.keys() )}' ) __UpperCamelCase : Any =processed_features[self.model_input_names[0]] __UpperCamelCase : Optional[Any] =( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowerCamelCase__ ) == 0: if return_attention_mask: __UpperCamelCase : str =[] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch __UpperCamelCase : Dict =required_input[0] if isinstance(lowerCamelCase__ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. __UpperCamelCase : int =0 while len(required_input[index] ) == 0: index += 1 if index < len(lowerCamelCase__ ): __UpperCamelCase : Any =required_input[index][0] if return_tensors is None: if is_tf_tensor(lowerCamelCase__ ): __UpperCamelCase : str ='tf' elif is_torch_tensor(lowerCamelCase__ ): __UpperCamelCase : Union[str, Any] ='pt' elif isinstance(lowerCamelCase__ , (int, float, list, tuple, np.ndarray) ): __UpperCamelCase : List[str] ='np' else: raise ValueError( f'type of {first_element} unknown: {type(lowerCamelCase__ )}. ' 'Should be one of a python, numpy, pytorch or tensorflow object.' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): __UpperCamelCase : Dict =to_numpy(lowerCamelCase__ ) else: __UpperCamelCase : List[Any] =[to_numpy(lowerCamelCase__ ) for v in value] # Convert padding_strategy in PaddingStrategy __UpperCamelCase : Dict =self._get_padding_strategies(padding=lowerCamelCase__ , max_length=lowerCamelCase__ ) __UpperCamelCase : int =processed_features[self.model_input_names[0]] __UpperCamelCase : int =len(lowerCamelCase__ ) if not all(len(lowerCamelCase__ ) == batch_size for v in processed_features.values() ): raise ValueError('Some items in the output dictionary have a different batch size than others.' ) __UpperCamelCase : Union[str, Any] =[] for i in range(lowerCamelCase__ ): __UpperCamelCase : Tuple ={k: v[i] for k, v in processed_features.items()} # truncation __UpperCamelCase : Tuple =self._truncate( lowerCamelCase__ , max_length=lowerCamelCase__ , pad_to_multiple_of=lowerCamelCase__ , truncation=lowerCamelCase__ , ) truncated_inputs.append(lowerCamelCase__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length __UpperCamelCase : Tuple =max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) __UpperCamelCase : List[Any] =PaddingStrategy.MAX_LENGTH __UpperCamelCase : Any ={} for i in range(lowerCamelCase__ ): # padding __UpperCamelCase : Union[str, Any] =self._pad( truncated_inputs[i] , max_length=lowerCamelCase__ , padding_strategy=lowerCamelCase__ , pad_to_multiple_of=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , ) for key, value in outputs.items(): if key not in batch_outputs: __UpperCamelCase : str =[] if value.dtype is np.dtype(np.floataa ): __UpperCamelCase : Tuple =value.astype(np.floataa ) batch_outputs[key].append(lowerCamelCase__ ) return BatchFeature(lowerCamelCase__ , tensor_type=lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = PaddingStrategy.DO_NOT_PAD , lowerCamelCase__ = None , lowerCamelCase__ = None , ): """simple docstring""" __UpperCamelCase : Dict =processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: __UpperCamelCase : List[str] =len(lowerCamelCase__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): __UpperCamelCase : List[Any] =((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of __UpperCamelCase : List[str] =padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowerCamelCase__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: __UpperCamelCase : Tuple =np.ones(len(lowerCamelCase__ ) , dtype=np.intaa ) if needs_to_be_padded: __UpperCamelCase : Any =max_length - len(lowerCamelCase__ ) if self.padding_side == "right": if return_attention_mask: __UpperCamelCase : Optional[Any] =np.pad( processed_features['attention_mask'] , (0, difference) ) __UpperCamelCase : List[str] =((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) __UpperCamelCase : Union[str, Any] =np.pad( lowerCamelCase__ , lowerCamelCase__ , 'constant' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: __UpperCamelCase : Any =np.pad( processed_features['attention_mask'] , (difference, 0) ) __UpperCamelCase : int =((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) __UpperCamelCase : str =np.pad( lowerCamelCase__ , lowerCamelCase__ , 'constant' , constant_values=self.padding_value ) else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return processed_features def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , ): """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('When setting ``truncation=True``, make sure that ``max_length`` is defined.' ) __UpperCamelCase : str =processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): __UpperCamelCase : Optional[int] =((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of __UpperCamelCase : Union[str, Any] =len(lowerCamelCase__ ) > max_length if needs_to_be_truncated: __UpperCamelCase : Dict =processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: __UpperCamelCase : Any =processed_features['attention_mask'][:max_length] return processed_features def __lowercase ( self , lowerCamelCase__=False , lowerCamelCase__=None ): """simple docstring""" if padding is not False: if padding is True: __UpperCamelCase : Optional[Any] =PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowerCamelCase__ , lowerCamelCase__ ): __UpperCamelCase : Optional[int] =PaddingStrategy(lowerCamelCase__ ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): __UpperCamelCase : List[str] =padding else: __UpperCamelCase : List[Any] =PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( f'When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined' ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( 'Asking to pad but the feature_extractor does not have a padding value. Please select a value to use' ' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.' ) return padding_strategy
71
A_ :Union[str, Any] = { 0: '''0''', 1: '''1''', 2: '''2''', 3: '''3''', 4: '''4''', 5: '''5''', 6: '''6''', 7: '''7''', 8: '''8''', 9: '''9''', 10: '''a''', 11: '''b''', 12: '''c''', 13: '''d''', 14: '''e''', 15: '''f''', } def A ( a_ ) -> str: assert type(a_ ) in (int, float) and decimal == int(a_ ) __UpperCamelCase : Union[str, Any] =int(a_ ) __UpperCamelCase : List[str] ='' __UpperCamelCase : Optional[Any] =False if decimal < 0: __UpperCamelCase : Tuple =True decimal *= -1 while decimal > 0: __UpperCamelCase , __UpperCamelCase : Optional[Any] =divmod(a_ ,16 ) __UpperCamelCase : Tuple =values[remainder] + hexadecimal __UpperCamelCase : Dict ='0x' + hexadecimal if negative: __UpperCamelCase : int ='-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
71
1
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def A ( a_ ,a_=0.999 ,a_="cosine" ,) -> Optional[Any]: if alpha_transform_type == "cosine": def alpha_bar_fn(a_ ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(a_ ): return math.exp(t * -12.0 ) else: raise ValueError(F'Unsupported alpha_tranform_type: {alpha_transform_type}' ) __UpperCamelCase : str =[] for i in range(a_ ): __UpperCamelCase : Optional[int] =i / num_diffusion_timesteps __UpperCamelCase : Optional[Any] =(i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(a_ ) / alpha_bar_fn(a_ ) ,a_ ) ) return torch.tensor(a_ ,dtype=torch.floataa ) class __A ( a , a ): """simple docstring""" UpperCamelCase__ : Tuple =[e.name for e in KarrasDiffusionSchedulers] UpperCamelCase__ : Optional[Any] =2 @register_to_config def __init__( self , lowerCamelCase__ = 1000 , lowerCamelCase__ = 0.00_085 , lowerCamelCase__ = 0.012 , lowerCamelCase__ = "linear" , lowerCamelCase__ = None , lowerCamelCase__ = "epsilon" , lowerCamelCase__ = "linspace" , lowerCamelCase__ = 0 , ): """simple docstring""" if trained_betas is not None: __UpperCamelCase : Optional[int] =torch.tensor(lowerCamelCase__ , dtype=torch.floataa ) elif beta_schedule == "linear": __UpperCamelCase : Tuple =torch.linspace(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. __UpperCamelCase : int =( torch.linspace(beta_start**0.5 , beta_end**0.5 , lowerCamelCase__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule __UpperCamelCase : Dict =betas_for_alpha_bar(lowerCamelCase__ ) else: raise NotImplementedError(f'{beta_schedule} does is not implemented for {self.__class__}' ) __UpperCamelCase : Dict =1.0 - self.betas __UpperCamelCase : List[Any] =torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=None ): """simple docstring""" if schedule_timesteps is None: __UpperCamelCase : Dict =self.timesteps __UpperCamelCase : List[Any] =(schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: __UpperCamelCase : List[str] =1 if len(lowerCamelCase__ ) > 1 else 0 else: __UpperCamelCase : int =timestep.cpu().item() if torch.is_tensor(lowerCamelCase__ ) else timestep __UpperCamelCase : Optional[Any] =self._index_counter[timestep_int] return indices[pos].item() @property def __lowercase ( self ): """simple docstring""" if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.index_for_timestep(lowerCamelCase__ ) if self.state_in_first_order: __UpperCamelCase : str =self.sigmas[step_index] else: __UpperCamelCase : Optional[int] =self.sigmas_interpol[step_index] __UpperCamelCase : Any =sample / ((sigma**2 + 1) ** 0.5) return sample def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , ): """simple docstring""" __UpperCamelCase : Dict =num_inference_steps __UpperCamelCase : Optional[int] =num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": __UpperCamelCase : Optional[int] =np.linspace(0 , num_train_timesteps - 1 , lowerCamelCase__ , dtype=lowerCamelCase__ )[::-1].copy() elif self.config.timestep_spacing == "leading": __UpperCamelCase : int =num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __UpperCamelCase : List[Any] =(np.arange(0 , lowerCamelCase__ ) * step_ratio).round()[::-1].copy().astype(lowerCamelCase__ ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": __UpperCamelCase : List[str] =num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __UpperCamelCase : Optional[int] =(np.arange(lowerCamelCase__ , 0 , -step_ratio )).round().copy().astype(lowerCamelCase__ ) timesteps -= 1 else: raise ValueError( f'{self.config.timestep_spacing} is not supported. Please make sure to choose one of \'linspace\', \'leading\' or \'trailing\'.' ) __UpperCamelCase : Any =np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) __UpperCamelCase : Optional[Any] =torch.from_numpy(np.log(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =np.interp(lowerCamelCase__ , np.arange(0 , len(lowerCamelCase__ ) ) , lowerCamelCase__ ) __UpperCamelCase : List[Any] =np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) __UpperCamelCase : Dict =torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ ) # interpolate sigmas __UpperCamelCase : Tuple =sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() __UpperCamelCase : int =torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) __UpperCamelCase : Optional[Any] =torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(lowerCamelCase__ ).startswith('mps' ): # mps does not support float64 __UpperCamelCase : Tuple =torch.from_numpy(lowerCamelCase__ ).to(lowerCamelCase__ , dtype=torch.floataa ) else: __UpperCamelCase : List[str] =torch.from_numpy(lowerCamelCase__ ).to(lowerCamelCase__ ) # interpolate timesteps __UpperCamelCase : Union[str, Any] =self.sigma_to_t(lowerCamelCase__ ).to(lowerCamelCase__ , dtype=timesteps.dtype ) __UpperCamelCase : List[Any] =torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() __UpperCamelCase : List[Any] =torch.cat([timesteps[:1], interleaved_timesteps] ) __UpperCamelCase : Optional[int] =None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter __UpperCamelCase : List[Any] =defaultdict(lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Dict =sigma.log() # get distribution __UpperCamelCase : Optional[int] =log_sigma - self.log_sigmas[:, None] # get sigmas range __UpperCamelCase : List[str] =dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) __UpperCamelCase : List[Any] =low_idx + 1 __UpperCamelCase : int =self.log_sigmas[low_idx] __UpperCamelCase : Tuple =self.log_sigmas[high_idx] # interpolate sigmas __UpperCamelCase : str =(low - log_sigma) / (low - high) __UpperCamelCase : List[Any] =w.clamp(0 , 1 ) # transform interpolation to time range __UpperCamelCase : List[str] =(1 - w) * low_idx + w * high_idx __UpperCamelCase : Dict =t.view(sigma.shape ) return t @property def __lowercase ( self ): """simple docstring""" return self.sample is None def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = True , ): """simple docstring""" __UpperCamelCase : int =self.index_for_timestep(lowerCamelCase__ ) # advance index counter by 1 __UpperCamelCase : str =timestep.cpu().item() if torch.is_tensor(lowerCamelCase__ ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: __UpperCamelCase : List[str] =self.sigmas[step_index] __UpperCamelCase : str =self.sigmas_interpol[step_index + 1] __UpperCamelCase : Any =self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method __UpperCamelCase : int =self.sigmas[step_index - 1] __UpperCamelCase : Tuple =self.sigmas_interpol[step_index] __UpperCamelCase : List[Any] =self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API __UpperCamelCase : Any =0 __UpperCamelCase : Tuple =sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": __UpperCamelCase : List[str] =sigma_hat if self.state_in_first_order else sigma_interpol __UpperCamelCase : Union[str, Any] =sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": __UpperCamelCase : Any =sigma_hat if self.state_in_first_order else sigma_interpol __UpperCamelCase : Optional[int] =model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError('prediction_type not implemented yet: sample' ) else: raise ValueError( f'prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`' ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order __UpperCamelCase : List[Any] =(sample - pred_original_sample) / sigma_hat # 3. delta timestep __UpperCamelCase : List[str] =sigma_interpol - sigma_hat # store for 2nd order step __UpperCamelCase : Tuple =sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order __UpperCamelCase : Optional[Any] =(sample - pred_original_sample) / sigma_interpol # 3. delta timestep __UpperCamelCase : Optional[Any] =sigma_next - sigma_hat __UpperCamelCase : List[Any] =self.sample __UpperCamelCase : int =None __UpperCamelCase : Union[str, Any] =sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : int =self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(lowerCamelCase__ ): # mps does not support float64 __UpperCamelCase : int =self.timesteps.to(original_samples.device , dtype=torch.floataa ) __UpperCamelCase : Union[str, Any] =timesteps.to(original_samples.device , dtype=torch.floataa ) else: __UpperCamelCase : Optional[Any] =self.timesteps.to(original_samples.device ) __UpperCamelCase : Tuple =timesteps.to(original_samples.device ) __UpperCamelCase : Any =[self.index_for_timestep(lowerCamelCase__ , lowerCamelCase__ ) for t in timesteps] __UpperCamelCase : List[Any] =sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): __UpperCamelCase : int =sigma.unsqueeze(-1 ) __UpperCamelCase : Dict =original_samples + noise * sigma return noisy_samples def __len__( self ): """simple docstring""" return self.config.num_train_timesteps
71
import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness A_ :List[str] = '''\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } ''' A_ :Any = '''\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). ''' A_ :Tuple = ''' Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric("code_eval") >>> test_cases = ["assert add(2,3)==5"] >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {\'pass@1\': 0.5, \'pass@2\': 1.0} ''' A_ :List[str] = ''' ################################################################################ !!!WARNING!!! ################################################################################ The "code_eval" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this with: >>> import os >>> os.environ["HF_ALLOW_CODE_EVAL"] = "1" ################################################################################\ ''' A_ :Tuple = '''The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): """simple docstring""" def __lowercase ( self ): """simple docstring""" return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' ) ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/openai/human-eval' , codebase_urls=['https://github.com/openai/human-eval'] , reference_urls=['https://github.com/openai/human-eval'] , license=_LICENSE , ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=[1, 10, 100] , lowerCamelCase__=4 , lowerCamelCase__=3.0 ): """simple docstring""" if os.getenv('HF_ALLOW_CODE_EVAL' , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError('This metric is currently not supported on Windows.' ) with ThreadPoolExecutor(max_workers=lowerCamelCase__ ) as executor: __UpperCamelCase : List[str] =[] __UpperCamelCase : Any =Counter() __UpperCamelCase : List[Any] =0 __UpperCamelCase : int =defaultdict(lowerCamelCase__ ) for task_id, (candidates, test_case) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): for candidate in candidates: __UpperCamelCase : str =candidate + '\n' + test_case __UpperCamelCase : Any =(test_program, timeout, task_id, completion_id[task_id]) __UpperCamelCase : Optional[Any] =executor.submit(lowerCamelCase__ , *lowerCamelCase__ ) futures.append(lowerCamelCase__ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(lowerCamelCase__ ): __UpperCamelCase : str =future.result() results[result["task_id"]].append((result['completion_id'], result) ) __UpperCamelCase , __UpperCamelCase : int =[], [] for result in results.values(): result.sort() __UpperCamelCase : str =[r[1]['passed'] for r in result] total.append(len(lowerCamelCase__ ) ) correct.append(sum(lowerCamelCase__ ) ) __UpperCamelCase : Optional[int] =np.array(lowerCamelCase__ ) __UpperCamelCase : List[str] =np.array(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =k __UpperCamelCase : List[Any] ={f'pass@{k}': estimate_pass_at_k(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def A ( a_ ,a_ ,a_ ) -> Optional[int]: def estimator(a_ ,a_ ,a_ ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 ,n + 1 ) ) if isinstance(a_ ,a_ ): __UpperCamelCase : Optional[int] =itertools.repeat(a_ ,len(a_ ) ) else: assert len(a_ ) == len(a_ ) __UpperCamelCase : List[Any] =iter(a_ ) return np.array([estimator(int(a_ ) ,int(a_ ) ,a_ ) for n, c in zip(a_ ,a_ )] )
71
1
import logging from transformers import PretrainedConfig A_ :Optional[int] = logging.getLogger(__name__) A_ :Any = { '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class __A ( a ): """simple docstring""" UpperCamelCase__ : Union[str, Any] ="""bertabs""" def __init__( self , lowerCamelCase__=30522 , lowerCamelCase__=512 , lowerCamelCase__=6 , lowerCamelCase__=512 , lowerCamelCase__=8 , lowerCamelCase__=512 , lowerCamelCase__=0.2 , lowerCamelCase__=6 , lowerCamelCase__=768 , lowerCamelCase__=8 , lowerCamelCase__=2048 , lowerCamelCase__=0.2 , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : str =vocab_size __UpperCamelCase : Any =max_pos __UpperCamelCase : List[Any] =enc_layers __UpperCamelCase : Dict =enc_hidden_size __UpperCamelCase : Union[str, Any] =enc_heads __UpperCamelCase : Any =enc_ff_size __UpperCamelCase : Optional[Any] =enc_dropout __UpperCamelCase : Any =dec_layers __UpperCamelCase : str =dec_hidden_size __UpperCamelCase : Optional[int] =dec_heads __UpperCamelCase : List[Any] =dec_ff_size __UpperCamelCase : Tuple =dec_dropout
71
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __A ( a , a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Optional[Any] =StableDiffusionDiffEditPipeline UpperCamelCase__ : str =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""height""", """width""", """image"""} | {"""image_latents"""} UpperCamelCase__ : Optional[Any] =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"""image"""} | {"""image_latents"""} UpperCamelCase__ : Dict =frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess UpperCamelCase__ : Any =frozenset([] ) def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Dict =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , ) __UpperCamelCase : List[str] =DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) __UpperCamelCase : Union[str, Any] =DDIMInverseScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_zero=lowerCamelCase__ , ) torch.manual_seed(0 ) __UpperCamelCase : Optional[int] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __UpperCamelCase : Tuple =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) __UpperCamelCase : Any =CLIPTextModel(lowerCamelCase__ ) __UpperCamelCase : int =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __UpperCamelCase : Union[str, Any] ={ 'unet': unet, 'scheduler': scheduler, 'inverse_scheduler': inverse_scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : int =floats_tensor((1, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Optional[int] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Dict ={ 'prompt': 'a dog and a newt', 'mask_image': mask, 'image_latents': latents, 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : Tuple =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : int =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : Optional[Any] =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : List[Any] =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Any =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'source_prompt': 'a cat and a frog', 'target_prompt': 'a dog and a newt', 'generator': generator, 'num_inference_steps': 2, 'num_maps_per_mask': 2, 'mask_encode_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : str =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : int =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : int =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'prompt': 'a cat and a frog', 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'decode_latents': True, 'output_type': 'numpy', } return inputs def __lowercase ( self ): """simple docstring""" if not hasattr(self.pipeline_class , '_optional_components' ): return __UpperCamelCase : Optional[Any] =self.get_dummy_components() __UpperCamelCase : List[str] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) __UpperCamelCase : Union[str, Any] =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe(**lowerCamelCase__ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : Tuple =self.pipeline_class.from_pretrained(lowerCamelCase__ ) pipe_loaded.to(lowerCamelCase__ ) pipe_loaded.set_progress_bar_config(disable=lowerCamelCase__ ) for optional_component in pipe._optional_components: self.assertTrue( getattr(lowerCamelCase__ , lowerCamelCase__ ) is None , f'`{optional_component}` did not stay set to None after loading.' , ) __UpperCamelCase : str =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe_loaded(**lowerCamelCase__ )[0] __UpperCamelCase : Tuple =np.abs(output - output_loaded ).max() self.assertLess(lowerCamelCase__ , 1E-4 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : int =self.get_dummy_mask_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe.generate_mask(**lowerCamelCase__ ) __UpperCamelCase : int =mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) __UpperCamelCase : Tuple =np.array([0] * 9 ) __UpperCamelCase : str =np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Optional[Any] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Dict =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : Optional[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : int =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def __lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=5E-3 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ='cpu' __UpperCamelCase : int =self.get_dummy_components() __UpperCamelCase : str ={'beta_start': 0.00_085, 'beta_end': 0.012, 'beta_schedule': 'scaled_linear'} __UpperCamelCase : str =DPMSolverMultistepScheduler(**lowerCamelCase__ ) __UpperCamelCase : Dict =DPMSolverMultistepInverseScheduler(**lowerCamelCase__ ) __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : str =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : List[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : Optional[Any] =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) @require_torch_gpu @slow class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def __lowercase ( cls ): """simple docstring""" __UpperCamelCase : Optional[int] =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png' ) __UpperCamelCase : Union[str, Any] =raw_image.convert('RGB' ).resize((768, 768) ) __UpperCamelCase : List[Any] =raw_image def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =torch.manual_seed(0 ) __UpperCamelCase : Dict =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : List[str] =DDIMScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : List[str] =DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : List[str] ='a bowl of fruit' __UpperCamelCase : Dict ='a bowl of pears' __UpperCamelCase : Tuple =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : int =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ ).latents __UpperCamelCase : Dict =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , output_type='numpy' , ).images[0] __UpperCamelCase : str =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =torch.manual_seed(0 ) __UpperCamelCase : List[Any] =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : Optional[Any] =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : Optional[int] =DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] ='a bowl of fruit' __UpperCamelCase : int ='a bowl of pears' __UpperCamelCase : str =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : List[str] =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ , num_inference_steps=25 , ).latents __UpperCamelCase : List[str] =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , num_inference_steps=25 , output_type='numpy' , ).images[0] __UpperCamelCase : Tuple =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1
71
1
import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters A_ :Optional[Any] = logging.get_logger(__name__) def A ( a_ ,a_ ,a_ ,a_=None ,a_=None ) -> int: # Recurse if needed if "." in tensor_name: __UpperCamelCase : Any =tensor_name.split('.' ) for split in splits[:-1]: __UpperCamelCase : Optional[Any] =getattr(a_ ,a_ ) if new_module is None: raise ValueError(F'{module} has no attribute {split}.' ) __UpperCamelCase : List[Any] =new_module __UpperCamelCase : List[str] =splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(F'{module} does not have a parameter or a buffer named {tensor_name}.' ) __UpperCamelCase : str =tensor_name in module._buffers __UpperCamelCase : Any =getattr(a_ ,a_ ) if old_value.device == torch.device('meta' ) and device not in ["meta", torch.device('meta' )] and value is None: raise ValueError(F'{tensor_name} is on the meta device, we need a `value` to put in on {device}.' ) __UpperCamelCase : str =False __UpperCamelCase : Any =False if is_buffer or not is_bitsandbytes_available(): __UpperCamelCase : str =False __UpperCamelCase : Union[str, Any] =False else: __UpperCamelCase : int =hasattr(bnb.nn ,'Params4bit' ) and isinstance(module._parameters[tensor_name] ,bnb.nn.Paramsabit ) __UpperCamelCase : Optional[Any] =isinstance(module._parameters[tensor_name] ,bnb.nn.IntaParams ) if is_abit or is_abit: __UpperCamelCase : Optional[Any] =module._parameters[tensor_name] if param.device.type != "cuda": if value is None: __UpperCamelCase : Tuple =old_value.to(a_ ) elif isinstance(a_ ,torch.Tensor ): __UpperCamelCase : Any =value.to('cpu' ) if value.dtype == torch.inta: __UpperCamelCase : Union[str, Any] =version.parse(importlib.metadata.version('bitsandbytes' ) ) > version.parse( '0.37.2' ) if not is_abit_serializable: raise ValueError( 'Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. ' 'Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`.' ) else: __UpperCamelCase : List[str] =torch.tensor(a_ ,device='cpu' ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls ,a_ ) and fpaa_statistics is None: __UpperCamelCase : Union[str, Any] =new_value.T __UpperCamelCase : Optional[int] =old_value.__dict__ if is_abit: __UpperCamelCase : Any =bnb.nn.IntaParams(a_ ,requires_grad=a_ ,**a_ ).to(a_ ) elif is_abit: __UpperCamelCase : List[str] =bnb.nn.Paramsabit(a_ ,requires_grad=a_ ,**a_ ).to(a_ ) __UpperCamelCase : Optional[Any] =new_value if fpaa_statistics is not None: setattr(module.weight ,'SCB' ,fpaa_statistics.to(a_ ) ) else: if value is None: __UpperCamelCase : Optional[Any] =old_value.to(a_ ) elif isinstance(a_ ,torch.Tensor ): __UpperCamelCase : Optional[int] =value.to(a_ ) else: __UpperCamelCase : Tuple =torch.tensor(a_ ,device=a_ ) if is_buffer: __UpperCamelCase : int =new_value else: __UpperCamelCase : List[str] =nn.Parameter(a_ ,requires_grad=old_value.requires_grad ) __UpperCamelCase : Optional[Any] =new_value def A ( a_ ,a_=None ,a_=None ,a_=None ,a_=False ) -> Tuple: for name, module in model.named_children(): if current_key_name is None: __UpperCamelCase : Any =[] current_key_name.append(a_ ) if (isinstance(a_ ,nn.Linear ) or isinstance(a_ ,a_ )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in '.'.join(a_ ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(a_ ,a_ ): __UpperCamelCase , __UpperCamelCase : Tuple =module.weight.shape else: __UpperCamelCase : int =module.in_features __UpperCamelCase : str =module.out_features if quantization_config.quantization_method() == "llm_int8": __UpperCamelCase : List[Any] =bnb.nn.LinearabitLt( a_ ,a_ ,module.bias is not None ,has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight ,threshold=quantization_config.llm_inta_threshold ,) __UpperCamelCase : Any =True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: __UpperCamelCase : int =bnb.nn.Linearabit( a_ ,a_ ,module.bias is not None ,quantization_config.bnb_abit_compute_dtype ,compress_statistics=quantization_config.bnb_abit_use_double_quant ,quant_type=quantization_config.bnb_abit_quant_type ,) __UpperCamelCase : Tuple =True # Store the module class in case we need to transpose the weight later __UpperCamelCase : Dict =type(a_ ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(a_ ) if len(list(module.children() ) ) > 0: __UpperCamelCase , __UpperCamelCase : Optional[Any] =_replace_with_bnb_linear( a_ ,a_ ,a_ ,a_ ,has_been_replaced=a_ ,) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def A ( a_ ,a_=None ,a_=None ,a_=None ) -> Optional[int]: __UpperCamelCase : Any =['lm_head'] if modules_to_not_convert is None else modules_to_not_convert __UpperCamelCase , __UpperCamelCase : Optional[Any] =_replace_with_bnb_linear( a_ ,a_ ,a_ ,a_ ) if not has_been_replaced: logger.warning( 'You are loading your model in 8bit or 4bit but no linear modules were found in your model.' ' Please double check your model architecture, or submit an issue on github if you think this is' ' a bug.' ) return model def A ( *a_ ,**a_ ) -> Tuple: warnings.warn( '`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead' ,a_ ,) return replace_with_bnb_linear(*a_ ,**a_ ) def A ( *a_ ,**a_ ) -> List[str]: warnings.warn( '`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead' ,a_ ,) return set_module_quantized_tensor_to_device(*a_ ,**a_ ) def A ( a_ ) -> Dict: __UpperCamelCase : str =deepcopy(a_ ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() __UpperCamelCase : Optional[int] =find_tied_parameters(a_ ) # For compatibility with Accelerate < 0.18 if isinstance(a_ ,a_ ): __UpperCamelCase : Optional[int] =sum(list(tied_params.values() ) ,[] ) + list(tied_params.keys() ) else: __UpperCamelCase : int =sum(a_ ,[] ) __UpperCamelCase : Optional[int] =len(a_ ) > 0 # Check if it is a base model __UpperCamelCase : Union[str, Any] =not hasattr(a_ ,model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head __UpperCamelCase : List[Any] =list(model.named_children() ) __UpperCamelCase : List[Any] =[list_modules[-1][0]] # add last module together with tied weights __UpperCamelCase : str =set(a_ ) - set(a_ ) __UpperCamelCase : Optional[Any] =list(set(a_ ) ) + list(a_ ) # remove ".weight" from the keys __UpperCamelCase : int =['.weight', '.bias'] __UpperCamelCase : Dict =[] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: __UpperCamelCase : Optional[int] =name.replace(a_ ,'' ) filtered_module_names.append(a_ ) return filtered_module_names
71
import random from .binary_exp_mod import bin_exp_mod def A ( a_ ,a_=1_000 ) -> Optional[Any]: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd __UpperCamelCase : List[Any] =n - 1 __UpperCamelCase : Dict =0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) __UpperCamelCase : Optional[Any] =0 while count < prec: __UpperCamelCase : Dict =random.randint(2 ,n - 1 ) __UpperCamelCase : Optional[Any] =bin_exp_mod(a_ ,a_ ,a_ ) if b != 1: __UpperCamelCase : List[str] =True for _ in range(a_ ): if b == n - 1: __UpperCamelCase : Tuple =False break __UpperCamelCase : Dict =b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": A_ :str = abs(int(input('''Enter bound : ''').strip())) print('''Here\'s the list of primes:''') print(''', '''.join(str(i) for i in range(n + 1) if is_prime_big(i)))
71
1
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __A : """simple docstring""" @staticmethod def __lowercase ( *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" pass @is_pipeline_test @require_vision class __A ( unittest.TestCase ): """simple docstring""" @require_torch def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , ) __UpperCamelCase : Dict =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) __UpperCamelCase : int =image_classifier(lowerCamelCase__ , candidate_labels=['a', 'b', 'c'] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(lowerCamelCase__ ) , [ [{'score': 0.333, 'label': 'a'}, {'score': 0.333, 'label': 'b'}, {'score': 0.333, 'label': 'c'}], [{'score': 0.333, 'label': 'a'}, {'score': 0.333, 'label': 'c'}, {'score': 0.333, 'label': 'b'}], ] , ) __UpperCamelCase : Tuple =image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], ] , ) @require_tf def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =pipeline( model='hf-internal-testing/tiny-random-clip-zero-shot-image-classification' , framework='tf' ) __UpperCamelCase : str =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) __UpperCamelCase : List[Any] =image_classifier(lowerCamelCase__ , candidate_labels=['a', 'b', 'c'] ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [{'score': 0.333, 'label': 'a'}, {'score': 0.333, 'label': 'b'}, {'score': 0.333, 'label': 'c'}] , ) __UpperCamelCase : Tuple =image_classifier([image] * 5 , candidate_labels=['A', 'B', 'C'] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], [ {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, {'score': 0.333, 'label': ANY(lowerCamelCase__ )}, ], ] , ) @slow @require_torch def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , ) # This is an image of 2 cats with remotes and no planes __UpperCamelCase : List[str] =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) __UpperCamelCase : int =image_classifier(lowerCamelCase__ , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ] , ) __UpperCamelCase : Optional[Any] =image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ], ] * 5 , ) @slow @require_tf def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =pipeline( task='zero-shot-image-classification' , model='openai/clip-vit-base-patch32' , framework='tf' ) # This is an image of 2 cats with remotes and no planes __UpperCamelCase : List[Any] =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) __UpperCamelCase : Optional[int] =image_classifier(lowerCamelCase__ , candidate_labels=['cat', 'plane', 'remote'] ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ] , ) __UpperCamelCase : Optional[Any] =image_classifier([image] * 5 , candidate_labels=['cat', 'plane', 'remote'] , batch_size=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , [ [ {'score': 0.511, 'label': 'remote'}, {'score': 0.485, 'label': 'cat'}, {'score': 0.004, 'label': 'plane'}, ], ] * 5 , )
71
from torch import nn class __A ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" super().__init__() __UpperCamelCase : Dict =class_size __UpperCamelCase : Any =embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) __UpperCamelCase : Any =nn.Linear(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : List[Any] =self.mlp(lowerCamelCase__ ) return logits
71
1
import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='| <pad> <unk> <s> </s> a b c d e f g h i j k'.split() __UpperCamelCase : List[str] =dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) __UpperCamelCase : Tuple ={ 'unk_token': '<unk>', 'bos_token': '<s>', 'eos_token': '</s>', } __UpperCamelCase : Any ={ 'feature_size': 1, 'padding_value': 0.0, 'sampling_rate': 16000, 'return_attention_mask': False, 'do_normalize': True, } __UpperCamelCase : Optional[Any] =tempfile.mkdtemp() __UpperCamelCase : List[str] =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase : Optional[Any] =os.path.join(self.tmpdirname , lowerCamelCase__ ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '\n' ) with open(self.feature_extraction_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '\n' ) # load decoder from hub __UpperCamelCase : Dict ='hf-internal-testing/ngram-beam-search-decoder' def __lowercase ( self , **lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Any =self.add_kwargs_tokens_map.copy() kwargs.update(lowerCamelCase__ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def __lowercase ( self , **lowerCamelCase__ ): """simple docstring""" return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def __lowercase ( self , **lowerCamelCase__ ): """simple docstring""" return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =self.get_tokenizer() __UpperCamelCase : Tuple =self.get_feature_extractor() __UpperCamelCase : Any =self.get_decoder() __UpperCamelCase : Tuple =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) __UpperCamelCase : Tuple =WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowerCamelCase__ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , lowerCamelCase__ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match __UpperCamelCase : List[str] =WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['xx'] ) with self.assertRaisesRegex(lowerCamelCase__ , 'include' ): WavaVecaProcessorWithLM( tokenizer=lowerCamelCase__ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.get_feature_extractor() __UpperCamelCase : List[str] =self.get_tokenizer() __UpperCamelCase : int =self.get_decoder() __UpperCamelCase : List[str] =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) __UpperCamelCase : Any =floats_list((3, 1000) ) __UpperCamelCase : Optional[int] =feature_extractor(lowerCamelCase__ , return_tensors='np' ) __UpperCamelCase : Optional[Any] =processor(lowerCamelCase__ , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =self.get_feature_extractor() __UpperCamelCase : Union[str, Any] =self.get_tokenizer() __UpperCamelCase : Dict =self.get_decoder() __UpperCamelCase : List[Any] =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) __UpperCamelCase : Dict ='This is a test string' __UpperCamelCase : Union[str, Any] =processor(text=lowerCamelCase__ ) __UpperCamelCase : int =tokenizer(lowerCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __lowercase ( self , lowerCamelCase__=(2, 10, 16) , lowerCamelCase__=77 ): """simple docstring""" np.random.seed(lowerCamelCase__ ) return np.random.rand(*lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =self.get_feature_extractor() __UpperCamelCase : List[Any] =self.get_tokenizer() __UpperCamelCase : Dict =self.get_decoder() __UpperCamelCase : List[Any] =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) __UpperCamelCase : List[str] =self._get_dummy_logits(shape=(10, 16) , seed=13 ) __UpperCamelCase : Optional[Any] =processor.decode(lowerCamelCase__ ) __UpperCamelCase : Optional[int] =decoder.decode_beams(lowerCamelCase__ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('</s> <s> </s>' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['fork'], ['spawn']] ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.get_feature_extractor() __UpperCamelCase : List[str] =self.get_tokenizer() __UpperCamelCase : Union[str, Any] =self.get_decoder() __UpperCamelCase : Tuple =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) __UpperCamelCase : List[str] =self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: __UpperCamelCase : Optional[Any] =processor.batch_decode(lowerCamelCase__ ) else: with get_context(lowerCamelCase__ ).Pool() as pool: __UpperCamelCase : Union[str, Any] =processor.batch_decode(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =list(lowerCamelCase__ ) with get_context('fork' ).Pool() as p: __UpperCamelCase : List[Any] =decoder.decode_beams_batch(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase : List[Any] =[], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(lowerCamelCase__ , decoded_processor.text ) self.assertListEqual(['<s> <s> </s>', '<s> <s> <s>'] , decoded_processor.text ) self.assertListEqual(lowerCamelCase__ , decoded_processor.logit_score ) self.assertListEqual(lowerCamelCase__ , decoded_processor.lm_score ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =self.get_feature_extractor() __UpperCamelCase : str =self.get_tokenizer() __UpperCamelCase : Optional[Any] =self.get_decoder() __UpperCamelCase : Optional[Any] =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) __UpperCamelCase : Optional[int] =self._get_dummy_logits() __UpperCamelCase : Dict =15 __UpperCamelCase : int =-20.0 __UpperCamelCase : Tuple =-4.0 __UpperCamelCase : Optional[int] =processor.batch_decode( lowerCamelCase__ , beam_width=lowerCamelCase__ , beam_prune_logp=lowerCamelCase__ , token_min_logp=lowerCamelCase__ , ) __UpperCamelCase : str =decoded_processor_out.text __UpperCamelCase : Tuple =list(lowerCamelCase__ ) with get_context('fork' ).Pool() as pool: __UpperCamelCase : List[str] =decoder.decode_beams_batch( lowerCamelCase__ , lowerCamelCase__ , beam_width=lowerCamelCase__ , beam_prune_logp=lowerCamelCase__ , token_min_logp=lowerCamelCase__ , ) __UpperCamelCase : Union[str, Any] =[d[0][0] for d in decoded_decoder_out] __UpperCamelCase : Dict =[d[0][2] for d in decoded_decoder_out] __UpperCamelCase : str =[d[0][3] for d in decoded_decoder_out] self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(['</s> <s> <s>', '<s> <s> <s>'] , lowerCamelCase__ ) self.assertTrue(np.array_equal(lowerCamelCase__ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , lowerCamelCase__ , atol=1E-3 ) ) self.assertTrue(np.array_equal(lowerCamelCase__ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9_474] , lowerCamelCase__ , atol=1E-3 ) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =self.get_feature_extractor() __UpperCamelCase : Union[str, Any] =self.get_tokenizer() __UpperCamelCase : Optional[int] =self.get_decoder() __UpperCamelCase : Optional[int] =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =self._get_dummy_logits() __UpperCamelCase : int =2.0 __UpperCamelCase : Union[str, Any] =5.0 __UpperCamelCase : List[str] =-20.0 __UpperCamelCase : str =True __UpperCamelCase : Any =processor.batch_decode( lowerCamelCase__ , alpha=lowerCamelCase__ , beta=lowerCamelCase__ , unk_score_offset=lowerCamelCase__ , lm_score_boundary=lowerCamelCase__ , ) __UpperCamelCase : int =decoded_processor_out.text __UpperCamelCase : Optional[int] =list(lowerCamelCase__ ) decoder.reset_params( alpha=lowerCamelCase__ , beta=lowerCamelCase__ , unk_score_offset=lowerCamelCase__ , lm_score_boundary=lowerCamelCase__ , ) with get_context('fork' ).Pool() as pool: __UpperCamelCase : str =decoder.decode_beams_batch( lowerCamelCase__ , lowerCamelCase__ , ) __UpperCamelCase : List[str] =[d[0][0] for d in decoded_decoder_out] self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(['<s> </s> <s> </s> </s>', '</s> </s> <s> </s> </s>'] , lowerCamelCase__ ) __UpperCamelCase : Any =processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm' ) __UpperCamelCase : Any =processor.decoder.model_container[processor.decoder._model_key] __UpperCamelCase : Any =Path(language_model._kenlm_model.path.decode('utf-8' ) ).parent.parent.absolute() __UpperCamelCase : List[Any] =os.listdir(lowerCamelCase__ ) __UpperCamelCase : str =['alphabet.json', 'language_model'] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =snapshot_download('hf-internal-testing/processor_with_lm' ) __UpperCamelCase : Any =WavaVecaProcessorWithLM.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : str =processor.decoder.model_container[processor.decoder._model_key] __UpperCamelCase : Optional[int] =Path(language_model._kenlm_model.path.decode('utf-8' ) ).parent.parent.absolute() __UpperCamelCase : Dict =os.listdir(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =os.listdir(lowerCamelCase__ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm' ) __UpperCamelCase : List[str] =AutoProcessor.from_pretrained('hf-internal-testing/processor_with_lm' ) __UpperCamelCase : int =floats_list((3, 1000) ) __UpperCamelCase : int =processor_wavaveca(lowerCamelCase__ , return_tensors='np' ) __UpperCamelCase : Optional[Any] =processor_auto(lowerCamelCase__ , return_tensors='np' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) __UpperCamelCase : Union[str, Any] =self._get_dummy_logits() __UpperCamelCase : Union[str, Any] =processor_wavaveca.batch_decode(lowerCamelCase__ ) __UpperCamelCase : str =processor_auto.batch_decode(lowerCamelCase__ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =self.get_feature_extractor() __UpperCamelCase : int =self.get_tokenizer() __UpperCamelCase : Any =self.get_decoder() __UpperCamelCase : int =WavaVecaProcessorWithLM(tokenizer=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , decoder=lowerCamelCase__ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='`processor` and `feature_extractor` model input names do not match' , ) @staticmethod def __lowercase ( lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : List[Any] =[d[key] for d in offsets] return retrieved_list def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm' ) __UpperCamelCase : Optional[int] =self._get_dummy_logits()[0] __UpperCamelCase : int =processor.decode(lowerCamelCase__ , output_word_offsets=lowerCamelCase__ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('text' in outputs ) self.assertTrue('word_offsets' in outputs ) self.assertTrue(isinstance(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(' '.join(self.get_from_offsets(outputs['word_offsets'] , 'word' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'] , 'word' ) , ['<s>', '<s>', '</s>'] ) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'] , 'start_offset' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'] , 'end_offset' ) , [1, 3, 5] ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =WavaVecaProcessorWithLM.from_pretrained('hf-internal-testing/processor_with_lm' ) __UpperCamelCase : Dict =self._get_dummy_logits() __UpperCamelCase : List[Any] =processor.batch_decode(lowerCamelCase__ , output_word_offsets=lowerCamelCase__ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('text' in outputs ) self.assertTrue('word_offsets' in outputs ) self.assertTrue(isinstance(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertListEqual( [' '.join(self.get_from_offsets(lowerCamelCase__ , 'word' ) ) for o in outputs['word_offsets']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'][0] , 'word' ) , ['<s>', '<s>', '</s>'] ) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'][0] , 'start_offset' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['word_offsets'][0] , 'end_offset' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def __lowercase ( self ): """simple docstring""" import torch __UpperCamelCase : List[str] =load_dataset('common_voice' , 'en' , split='train' , streaming=lowerCamelCase__ ) __UpperCamelCase : Any =ds.cast_column('audio' , datasets.Audio(sampling_rate=16000 ) ) __UpperCamelCase : Any =iter(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =next(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =AutoProcessor.from_pretrained('patrickvonplaten/wav2vec2-base-100h-with-lm' ) __UpperCamelCase : Optional[int] =WavaVecaForCTC.from_pretrained('patrickvonplaten/wav2vec2-base-100h-with-lm' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train __UpperCamelCase : Optional[Any] =processor(sample['audio']['array'] , return_tensors='pt' ).input_values with torch.no_grad(): __UpperCamelCase : Any =model(lowerCamelCase__ ).logits.cpu().numpy() __UpperCamelCase : Union[str, Any] =processor.decode(logits[0] , output_word_offsets=lowerCamelCase__ ) __UpperCamelCase : List[Any] =model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate __UpperCamelCase : Optional[Any] =[ { 'start_time': d['start_offset'] * time_offset, 'end_time': d['end_offset'] * time_offset, 'word': d['word'], } for d in output['word_offsets'] ] __UpperCamelCase : int ='WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL' # output words self.assertEqual(' '.join(self.get_from_offsets(lowerCamelCase__ , 'word' ) ) , lowerCamelCase__ ) self.assertEqual(' '.join(self.get_from_offsets(lowerCamelCase__ , 'word' ) ) , output.text ) # output times __UpperCamelCase : Any =torch.tensor(self.get_from_offsets(lowerCamelCase__ , 'start_time' ) ) __UpperCamelCase : List[Any] =torch.tensor(self.get_from_offsets(lowerCamelCase__ , 'end_time' ) ) # fmt: off __UpperCamelCase : Union[str, Any] =torch.tensor([1.4_199, 1.6_599, 2.2_599, 3.0, 3.24, 3.5_999, 3.7_999, 4.0_999, 4.26, 4.94, 5.28, 5.6_599, 5.78, 5.94, 6.32, 6.5_399, 6.6_599] ) __UpperCamelCase : int =torch.tensor([1.5_399, 1.8_999, 2.9, 3.16, 3.5_399, 3.72, 4.0_199, 4.1_799, 4.76, 5.1_599, 5.5_599, 5.6_999, 5.86, 6.1_999, 6.38, 6.6_199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=0.01 ) ) self.assertTrue(torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=0.01 ) )
71
def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square(a_ ,a_ ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __UpperCamelCase : Optional[int] =update_area_of_max_square(a_ ,col + 1 ) __UpperCamelCase : List[str] =update_area_of_max_square(row + 1 ,col + 1 ) __UpperCamelCase : List[Any] =update_area_of_max_square(row + 1 ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : Dict =max(largest_square_area[0] ,a_ ) return sub_problem_sol else: return 0 __UpperCamelCase : Union[str, Any] =[0] update_area_of_max_square(0 ,0 ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square_using_dp_array( a_ ,a_ ,a_ ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __UpperCamelCase : Tuple =update_area_of_max_square_using_dp_array(a_ ,col + 1 ,a_ ) __UpperCamelCase : Optional[int] =update_area_of_max_square_using_dp_array(row + 1 ,col + 1 ,a_ ) __UpperCamelCase : Any =update_area_of_max_square_using_dp_array(row + 1 ,a_ ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : str =max(largest_square_area[0] ,a_ ) __UpperCamelCase : Any =sub_problem_sol return sub_problem_sol else: return 0 __UpperCamelCase : Tuple =[0] __UpperCamelCase : List[Any] =[[-1] * cols for _ in range(a_ )] update_area_of_max_square_using_dp_array(0 ,0 ,a_ ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Dict =[[0] * (cols + 1) for _ in range(rows + 1 )] __UpperCamelCase : int =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Optional[Any] =dp_array[row][col + 1] __UpperCamelCase : int =dp_array[row + 1][col + 1] __UpperCamelCase : Tuple =dp_array[row + 1][col] if mat[row][col] == 1: __UpperCamelCase : Tuple =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Any =max(dp_array[row][col] ,a_ ) else: __UpperCamelCase : Dict =0 return largest_square_area def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Any =[0] * (cols + 1) __UpperCamelCase : List[Any] =[0] * (cols + 1) __UpperCamelCase : Tuple =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Any =current_row[col + 1] __UpperCamelCase : Optional[Any] =next_row[col + 1] __UpperCamelCase : Union[str, Any] =next_row[col] if mat[row][col] == 1: __UpperCamelCase : Any =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Optional[int] =max(current_row[col] ,a_ ) else: __UpperCamelCase : List[str] =0 __UpperCamelCase : Optional[Any] =current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
71
1
from ..utils import DummyObject, requires_backends class __A ( metaclass=a ): """simple docstring""" UpperCamelCase__ : Union[str, Any] =["""note_seq"""] def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" requires_backends(self , ['note_seq'] ) @classmethod def __lowercase ( cls , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" requires_backends(cls , ['note_seq'] ) @classmethod def __lowercase ( cls , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" requires_backends(cls , ['note_seq'] )
71
def A ( a_ ) -> int: __UpperCamelCase : Any =len(a_ ) while cur > 1: # Find the maximum number in arr __UpperCamelCase : Any =arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi __UpperCamelCase : Any =arr[mi::-1] + arr[mi + 1 : len(a_ )] # Reverse whole list __UpperCamelCase : str =arr[cur - 1 :: -1] + arr[cur : len(a_ )] cur -= 1 return arr if __name__ == "__main__": A_ :Dict = input('''Enter numbers separated by a comma:\n''').strip() A_ :Any = [int(item) for item in user_input.split(''',''')] print(pancake_sort(unsorted))
71
1
def A ( a_ = 200 ) -> int: __UpperCamelCase : Optional[int] =[1, 2, 5, 10, 20, 50, 100, 200] __UpperCamelCase : List[str] =[0] * (pence + 1) __UpperCamelCase : Any =1 # base case: 1 way to make 0 pence for coin in coins: for i in range(a_ ,pence + 1 ,1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
71
import random def A ( a_ ,a_ ,a_ = False ) -> dict: __UpperCamelCase : dict ={i: [] for i in range(a_ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(a_ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(a_ ): for j in range(i + 1 ,a_ ): if random.random() < probability: graph[i].append(a_ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(a_ ) return graph def A ( a_ ) -> dict: return { i: [j for j in range(a_ ) if i != j] for i in range(a_ ) } if __name__ == "__main__": import doctest doctest.testmod()
71
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A_ :Tuple = { '''configuration_altclip''': [ '''ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AltCLIPConfig''', '''AltCLIPTextConfig''', '''AltCLIPVisionConfig''', ], '''processing_altclip''': ['''AltCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Optional[Any] = [ '''ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AltCLIPPreTrainedModel''', '''AltCLIPModel''', '''AltCLIPTextModel''', '''AltCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys A_ :List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
71
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[str] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModel.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =AutoModel.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForPreTraining.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =AutoModelForPreTraining.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Tuple =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Tuple =TFAutoModelForCausalLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =AutoModelForCausalLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =AutoModelForCausalLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : int =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Dict =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Tuple =TFAutoModelForMaskedLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =AutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Optional[Any] =AutoModelForMaskedLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[str] =TFAutoModelForSeqaSeqLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Dict =AutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[Any] =AutoModelForSeqaSeqLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : str =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =TFAutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =AutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =TFAutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =AutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) __UpperCamelCase : str =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) __UpperCamelCase : int =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 )
71
1
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import Callable, Dict, List, Tuple import timm import torch import torch.nn as nn from classy_vision.models.regnet import RegNet, RegNetParams, RegNetYaagf, RegNetYaagf, RegNetYaaagf from huggingface_hub import cached_download, hf_hub_url from torch import Tensor from vissl.models.model_helpers import get_trunk_forward_outputs from transformers import AutoImageProcessor, RegNetConfig, RegNetForImageClassification, RegNetModel from transformers.utils import logging logging.set_verbosity_info() A_ :Tuple = logging.get_logger() @dataclass class __A : """simple docstring""" UpperCamelCase__ : nn.Module UpperCamelCase__ : List[nn.Module] =field(default_factory=a ) UpperCamelCase__ : list =field(default_factory=a ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Dict =len(list(m.modules() ) ) == 1 or isinstance(lowerCamelCase__ , nn.Convad ) or isinstance(lowerCamelCase__ , nn.BatchNormad ) if has_not_submodules: self.traced.append(lowerCamelCase__ ) def __call__( self , lowerCamelCase__ ): """simple docstring""" for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(lowerCamelCase__ ) [x.remove() for x in self.handles] return self @property def __lowercase ( self ): """simple docstring""" return list(filter(lambda lowerCamelCase__ : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : nn.Module UpperCamelCase__ : nn.Module UpperCamelCase__ : int =1 UpperCamelCase__ : List =field(default_factory=a ) UpperCamelCase__ : List =field(default_factory=a ) UpperCamelCase__ : bool =True def __call__( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Any =Tracker(self.dest )(lowerCamelCase__ ).parametrized __UpperCamelCase : Dict =Tracker(self.src )(lowerCamelCase__ ).parametrized __UpperCamelCase : Union[str, Any] =list(filter(lambda lowerCamelCase__ : type(lowerCamelCase__ ) not in self.src_skip , lowerCamelCase__ ) ) __UpperCamelCase : Tuple =list(filter(lambda lowerCamelCase__ : type(lowerCamelCase__ ) not in self.dest_skip , lowerCamelCase__ ) ) if len(lowerCamelCase__ ) != len(lowerCamelCase__ ) and self.raise_if_mismatch: raise Exception( f'Numbers of operations are different. Source module has {len(lowerCamelCase__ )} operations while' f' destination module has {len(lowerCamelCase__ )}.' ) for dest_m, src_m in zip(lowerCamelCase__ , lowerCamelCase__ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(f'Transfered from={src_m} to={dest_m}' ) class __A ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase__ ): """simple docstring""" super().__init__() __UpperCamelCase : List[Tuple[str, nn.Module]] =[] # - get the stem feature_blocks.append(('conv1', model.stem) ) # - get all the feature blocks for k, v in model.trunk_output.named_children(): assert k.startswith('block' ), f'Unexpected layer name {k}' __UpperCamelCase : Dict =len(lowerCamelCase__ ) + 1 feature_blocks.append((f'res{block_index}', v) ) __UpperCamelCase : Any =nn.ModuleDict(lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" return get_trunk_forward_outputs( lowerCamelCase__ , out_feat_keys=lowerCamelCase__ , feature_blocks=self._feature_blocks , ) class __A ( a ): """simple docstring""" def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : List[str] =x.split('-' ) return x_split[0] + x_split[1] + "_" + "".join(x_split[2:] ) def __getitem__( self , lowerCamelCase__ ): """simple docstring""" if x not in self: __UpperCamelCase : Any =self.convert_name_to_timm(lowerCamelCase__ ) __UpperCamelCase : Optional[int] =partial(lambda: (timm.create_model(lowerCamelCase__ , pretrained=lowerCamelCase__ ).eval(), None) ) else: __UpperCamelCase : Dict =super().__getitem__(lowerCamelCase__ ) return val class __A ( a ): """simple docstring""" def __getitem__( self , lowerCamelCase__ ): """simple docstring""" if "seer" in x and "in1k" not in x: __UpperCamelCase : str =RegNetModel else: __UpperCamelCase : Union[str, Any] =RegNetForImageClassification return val def A ( a_ ,a_ ,a_ ) -> Any: for from_key, to_key in keys: __UpperCamelCase : str =from_state_dict[from_key].clone() print(F'Copied key={from_key} to={to_key}' ) return to_state_dict def A ( a_ ,a_ ,a_ ,a_ ,a_ ,a_ = True ,) -> Union[str, Any]: print(F'Converting {name}...' ) with torch.no_grad(): __UpperCamelCase , __UpperCamelCase : str =from_model_func() __UpperCamelCase : Dict =our_model_func(a_ ).eval() __UpperCamelCase : List[Any] =ModuleTransfer(src=a_ ,dest=a_ ,raise_if_mismatch=a_ ) __UpperCamelCase : List[str] =torch.randn((1, 3, 224, 224) ) module_transfer(a_ ) if from_state_dict is not None: __UpperCamelCase : Dict =[] # for seer - in1k finetuned we have to manually copy the head if "seer" in name and "in1k" in name: __UpperCamelCase : Optional[Any] =[('0.clf.0.weight', 'classifier.1.weight'), ('0.clf.0.bias', 'classifier.1.bias')] __UpperCamelCase : List[str] =manually_copy_vissl_head(a_ ,our_model.state_dict() ,a_ ) our_model.load_state_dict(a_ ) __UpperCamelCase : Optional[Any] =our_model(a_ ,output_hidden_states=a_ ) __UpperCamelCase : Union[str, Any] =( our_outputs.logits if isinstance(a_ ,a_ ) else our_outputs.last_hidden_state ) __UpperCamelCase : int =from_model(a_ ) __UpperCamelCase : Optional[Any] =from_output[-1] if type(a_ ) is list else from_output # now since I don't want to use any config files, vissl seer model doesn't actually have an head, so let's just check the last hidden state if "seer" in name and "in1k" in name: __UpperCamelCase : Optional[Any] =our_outputs.hidden_states[-1] assert torch.allclose(a_ ,a_ ), "The model logits don't match the original one." if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / name ,commit_message='Add model' ,use_temp_dir=a_ ,) __UpperCamelCase : List[Any] =224 if 'seer' not in name else 384 # we can use the convnext one __UpperCamelCase : List[str] =AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ,size=a_ ) image_processor.push_to_hub( repo_path_or_name=save_directory / name ,commit_message='Add image processor' ,use_temp_dir=a_ ,) print(F'Pushed {name}' ) def A ( a_ ,a_ = None ,a_ = True ) -> Union[str, Any]: __UpperCamelCase : Optional[Any] ='imagenet-1k-id2label.json' __UpperCamelCase : str =1_000 __UpperCamelCase : int =(1, num_labels) __UpperCamelCase : Tuple ='huggingface/label-files' __UpperCamelCase : Tuple =num_labels __UpperCamelCase : Dict =json.load(open(cached_download(hf_hub_url(a_ ,a_ ,repo_type='dataset' ) ) ,'r' ) ) __UpperCamelCase : Any ={int(a_ ): v for k, v in idalabel.items()} __UpperCamelCase : Optional[Any] =idalabel __UpperCamelCase : Tuple ={v: k for k, v in idalabel.items()} __UpperCamelCase : Tuple =partial(a_ ,num_labels=a_ ,idalabel=a_ ,labelaid=a_ ) __UpperCamelCase : Optional[Any] ={ 'regnet-x-002': ImageNetPreTrainedConfig( depths=[1, 1, 4, 7] ,hidden_sizes=[24, 56, 152, 368] ,groups_width=8 ,layer_type='x' ), 'regnet-x-004': ImageNetPreTrainedConfig( depths=[1, 2, 7, 12] ,hidden_sizes=[32, 64, 160, 384] ,groups_width=16 ,layer_type='x' ), 'regnet-x-006': ImageNetPreTrainedConfig( depths=[1, 3, 5, 7] ,hidden_sizes=[48, 96, 240, 528] ,groups_width=24 ,layer_type='x' ), 'regnet-x-008': ImageNetPreTrainedConfig( depths=[1, 3, 7, 5] ,hidden_sizes=[64, 128, 288, 672] ,groups_width=16 ,layer_type='x' ), 'regnet-x-016': ImageNetPreTrainedConfig( depths=[2, 4, 10, 2] ,hidden_sizes=[72, 168, 408, 912] ,groups_width=24 ,layer_type='x' ), 'regnet-x-032': ImageNetPreTrainedConfig( depths=[2, 6, 15, 2] ,hidden_sizes=[96, 192, 432, 1_008] ,groups_width=48 ,layer_type='x' ), 'regnet-x-040': ImageNetPreTrainedConfig( depths=[2, 5, 14, 2] ,hidden_sizes=[80, 240, 560, 1_360] ,groups_width=40 ,layer_type='x' ), 'regnet-x-064': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] ,hidden_sizes=[168, 392, 784, 1_624] ,groups_width=56 ,layer_type='x' ), 'regnet-x-080': ImageNetPreTrainedConfig( depths=[2, 5, 15, 1] ,hidden_sizes=[80, 240, 720, 1_920] ,groups_width=120 ,layer_type='x' ), 'regnet-x-120': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] ,hidden_sizes=[224, 448, 896, 2_240] ,groups_width=112 ,layer_type='x' ), 'regnet-x-160': ImageNetPreTrainedConfig( depths=[2, 6, 13, 1] ,hidden_sizes=[256, 512, 896, 2_048] ,groups_width=128 ,layer_type='x' ), 'regnet-x-320': ImageNetPreTrainedConfig( depths=[2, 7, 13, 1] ,hidden_sizes=[336, 672, 1_344, 2_520] ,groups_width=168 ,layer_type='x' ), # y variant 'regnet-y-002': ImageNetPreTrainedConfig(depths=[1, 1, 4, 7] ,hidden_sizes=[24, 56, 152, 368] ,groups_width=8 ), 'regnet-y-004': ImageNetPreTrainedConfig( depths=[1, 3, 6, 6] ,hidden_sizes=[48, 104, 208, 440] ,groups_width=8 ), 'regnet-y-006': ImageNetPreTrainedConfig( depths=[1, 3, 7, 4] ,hidden_sizes=[48, 112, 256, 608] ,groups_width=16 ), 'regnet-y-008': ImageNetPreTrainedConfig( depths=[1, 3, 8, 2] ,hidden_sizes=[64, 128, 320, 768] ,groups_width=16 ), 'regnet-y-016': ImageNetPreTrainedConfig( depths=[2, 6, 17, 2] ,hidden_sizes=[48, 120, 336, 888] ,groups_width=24 ), 'regnet-y-032': ImageNetPreTrainedConfig( depths=[2, 5, 13, 1] ,hidden_sizes=[72, 216, 576, 1_512] ,groups_width=24 ), 'regnet-y-040': ImageNetPreTrainedConfig( depths=[2, 6, 12, 2] ,hidden_sizes=[128, 192, 512, 1_088] ,groups_width=64 ), 'regnet-y-064': ImageNetPreTrainedConfig( depths=[2, 7, 14, 2] ,hidden_sizes=[144, 288, 576, 1_296] ,groups_width=72 ), 'regnet-y-080': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] ,hidden_sizes=[168, 448, 896, 2_016] ,groups_width=56 ), 'regnet-y-120': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] ,hidden_sizes=[224, 448, 896, 2_240] ,groups_width=112 ), 'regnet-y-160': ImageNetPreTrainedConfig( depths=[2, 4, 11, 1] ,hidden_sizes=[224, 448, 1_232, 3_024] ,groups_width=112 ), 'regnet-y-320': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] ,hidden_sizes=[232, 696, 1_392, 3_712] ,groups_width=232 ), # models created by SEER -> https://arxiv.org/abs/2202.08360 'regnet-y-320-seer': RegNetConfig(depths=[2, 5, 12, 1] ,hidden_sizes=[232, 696, 1_392, 3_712] ,groups_width=232 ), 'regnet-y-640-seer': RegNetConfig(depths=[2, 5, 12, 1] ,hidden_sizes=[328, 984, 1_968, 4_920] ,groups_width=328 ), 'regnet-y-1280-seer': RegNetConfig( depths=[2, 7, 17, 1] ,hidden_sizes=[528, 1_056, 2_904, 7_392] ,groups_width=264 ), 'regnet-y-2560-seer': RegNetConfig( depths=[3, 7, 16, 1] ,hidden_sizes=[640, 1_696, 2_544, 5_088] ,groups_width=640 ), 'regnet-y-10b-seer': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] ,hidden_sizes=[2_020, 4_040, 11_110, 28_280] ,groups_width=1_010 ), # finetuned on imagenet 'regnet-y-320-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] ,hidden_sizes=[232, 696, 1_392, 3_712] ,groups_width=232 ), 'regnet-y-640-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] ,hidden_sizes=[328, 984, 1_968, 4_920] ,groups_width=328 ), 'regnet-y-1280-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] ,hidden_sizes=[528, 1_056, 2_904, 7_392] ,groups_width=264 ), 'regnet-y-2560-seer-in1k': ImageNetPreTrainedConfig( depths=[3, 7, 16, 1] ,hidden_sizes=[640, 1_696, 2_544, 5_088] ,groups_width=640 ), 'regnet-y-10b-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] ,hidden_sizes=[2_020, 4_040, 11_110, 28_280] ,groups_width=1_010 ), } __UpperCamelCase : Optional[int] =NameToOurModelFuncMap() __UpperCamelCase : Optional[Any] =NameToFromModelFuncMap() # add seer weights logic def load_using_classy_vision(a_ ,a_ ) -> Tuple[nn.Module, Dict]: __UpperCamelCase : Dict =torch.hub.load_state_dict_from_url(a_ ,model_dir=str(a_ ) ,map_location='cpu' ) __UpperCamelCase : List[Any] =model_func() # check if we have a head, if yes add it __UpperCamelCase : str =files['classy_state_dict']['base_model']['model'] __UpperCamelCase : Any =model_state_dict['trunk'] model.load_state_dict(a_ ) return model.eval(), model_state_dict["heads"] # pretrained __UpperCamelCase : Union[str, Any] =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet32d/seer_regnet32gf_model_iteration244000.torch' ,lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) ,) __UpperCamelCase : Optional[int] =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet64/seer_regnet64gf_model_final_checkpoint_phase0.torch' ,lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) ,) __UpperCamelCase : str =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/swav_ig1b_regnet128Gf_cnstant_bs32_node16_sinkhorn10_proto16k_syncBN64_warmup8k/model_final_checkpoint_phase0.torch' ,lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) ,) __UpperCamelCase : Optional[int] =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet10B/model_iteration124500_conso.torch' ,lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 ,group_width=1_010 ,w_a=1_744 ,w_a=620.83 ,w_m=2.52 ) ) ) ,) # IN1K finetuned __UpperCamelCase : Union[str, Any] =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet32_finetuned_in1k_model_final_checkpoint_phase78.torch' ,lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) ,) __UpperCamelCase : Optional[int] =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet64_finetuned_in1k_model_final_checkpoint_phase78.torch' ,lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) ,) __UpperCamelCase : Union[str, Any] =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet128_finetuned_in1k_model_final_checkpoint_phase78.torch' ,lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) ,) __UpperCamelCase : str =partial( a_ ,'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_10b_finetuned_in1k_model_phase28_conso.torch' ,lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 ,group_width=1_010 ,w_a=1_744 ,w_a=620.83 ,w_m=2.52 ) ) ) ,) if model_name: convert_weight_and_push( a_ ,names_to_from_model_map[model_name] ,names_to_ours_model_map[model_name] ,names_to_config[model_name] ,a_ ,a_ ,) else: for model_name, config in names_to_config.items(): convert_weight_and_push( a_ ,names_to_from_model_map[model_name] ,names_to_ours_model_map[model_name] ,a_ ,a_ ,a_ ,) return config, expected_shape if __name__ == "__main__": A_ :Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default=None, type=str, help=( '''The name of the model you wish to convert, it must be one of the supported regnet* architecture,''' ''' currently: regnetx-*, regnety-*. If `None`, all of them will the converted.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=Path, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', default=True, type=bool, required=False, help='''If True, push model and image processor to the hub.''', ) A_ :List[Any] = parser.parse_args() A_ :Path = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
71
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ :Tuple = { '''configuration_x_clip''': [ '''XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XCLIPConfig''', '''XCLIPTextConfig''', '''XCLIPVisionConfig''', ], '''processing_x_clip''': ['''XCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Union[str, Any] = [ '''XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XCLIPModel''', '''XCLIPPreTrainedModel''', '''XCLIPTextModel''', '''XCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys A_ :Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
71
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available A_ :Optional[int] = { '''configuration_longt5''': ['''LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongT5Config''', '''LongT5OnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :str = [ '''LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongT5EncoderModel''', '''LongT5ForConditionalGeneration''', '''LongT5Model''', '''LongT5PreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :str = [ '''FlaxLongT5ForConditionalGeneration''', '''FlaxLongT5Model''', '''FlaxLongT5PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys A_ :Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
71
from ...configuration_utils import PretrainedConfig from ...utils import logging A_ :Any = logging.get_logger(__name__) A_ :int = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[int] ="""vit_msn""" def __init__( self , lowerCamelCase__=768 , lowerCamelCase__=12 , lowerCamelCase__=12 , lowerCamelCase__=3072 , lowerCamelCase__="gelu" , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.02 , lowerCamelCase__=1E-06 , lowerCamelCase__=224 , lowerCamelCase__=16 , lowerCamelCase__=3 , lowerCamelCase__=True , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : int =hidden_size __UpperCamelCase : List[Any] =num_hidden_layers __UpperCamelCase : Union[str, Any] =num_attention_heads __UpperCamelCase : List[str] =intermediate_size __UpperCamelCase : Union[str, Any] =hidden_act __UpperCamelCase : str =hidden_dropout_prob __UpperCamelCase : Union[str, Any] =attention_probs_dropout_prob __UpperCamelCase : Union[str, Any] =initializer_range __UpperCamelCase : Tuple =layer_norm_eps __UpperCamelCase : Optional[Any] =image_size __UpperCamelCase : Optional[int] =patch_size __UpperCamelCase : Any =num_channels __UpperCamelCase : str =qkv_bias
71
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType A_ :Any = logging.get_logger(__name__) A_ :List[str] = { '''openai/imagegpt-small''': '''''', '''openai/imagegpt-medium''': '''''', '''openai/imagegpt-large''': '''''', } class __A ( a ): """simple docstring""" UpperCamelCase__ : List[str] ="""imagegpt""" UpperCamelCase__ : int =["""past_key_values"""] UpperCamelCase__ : Optional[int] ={ """hidden_size""": """n_embd""", """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCamelCase__=512 + 1 , lowerCamelCase__=32 * 32 , lowerCamelCase__=512 , lowerCamelCase__=24 , lowerCamelCase__=8 , lowerCamelCase__=None , lowerCamelCase__="quick_gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1E-5 , lowerCamelCase__=0.02 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=False , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Any =vocab_size __UpperCamelCase : Optional[Any] =n_positions __UpperCamelCase : List[Any] =n_embd __UpperCamelCase : Optional[Any] =n_layer __UpperCamelCase : Dict =n_head __UpperCamelCase : Union[str, Any] =n_inner __UpperCamelCase : Union[str, Any] =activation_function __UpperCamelCase : List[Any] =resid_pdrop __UpperCamelCase : int =embd_pdrop __UpperCamelCase : Tuple =attn_pdrop __UpperCamelCase : Optional[Any] =layer_norm_epsilon __UpperCamelCase : Tuple =initializer_range __UpperCamelCase : Any =scale_attn_weights __UpperCamelCase : int =use_cache __UpperCamelCase : Union[str, Any] =scale_attn_by_inverse_layer_idx __UpperCamelCase : int =reorder_and_upcast_attn __UpperCamelCase : List[str] =tie_word_embeddings super().__init__(tie_word_embeddings=lowerCamelCase__ , **lowerCamelCase__ ) class __A ( a ): """simple docstring""" @property def __lowercase ( self ): """simple docstring""" return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ] ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = 1 , lowerCamelCase__ = -1 , lowerCamelCase__ = False , lowerCamelCase__ = None , lowerCamelCase__ = 3 , lowerCamelCase__ = 32 , lowerCamelCase__ = 32 , ): """simple docstring""" __UpperCamelCase : Optional[Any] =self._generate_dummy_images(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Any =dict(preprocessor(images=lowerCamelCase__ , return_tensors=lowerCamelCase__ ) ) return inputs
71
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __A ( a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : str =DDIMPipeline UpperCamelCase__ : List[Any] =UNCONDITIONAL_IMAGE_GENERATION_PARAMS UpperCamelCase__ : Tuple =PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } UpperCamelCase__ : Tuple =UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS UpperCamelCase__ : Any =False def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Optional[int] =UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) __UpperCamelCase : int =DDIMScheduler() __UpperCamelCase : Optional[int] ={'unet': unet, 'scheduler': scheduler} return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : str =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Optional[int] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Tuple ={ 'batch_size': 1, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='cpu' __UpperCamelCase : Optional[Any] =self.get_dummy_components() __UpperCamelCase : Tuple =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : int =pipe(**lowerCamelCase__ ).images __UpperCamelCase : Dict =image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) __UpperCamelCase : Tuple =np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) __UpperCamelCase : Tuple =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def __lowercase ( self ): """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str ='google/ddpm-cifar10-32' __UpperCamelCase : str =UNetaDModel.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =DDIMScheduler() __UpperCamelCase : List[Any] =DDIMPipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) ddim.to(lowerCamelCase__ ) ddim.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[int] =torch.manual_seed(0 ) __UpperCamelCase : List[str] =ddim(generator=lowerCamelCase__ , eta=0.0 , output_type='numpy' ).images __UpperCamelCase : Union[str, Any] =image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase : str =np.array([0.1_723, 0.1_617, 0.1_600, 0.1_626, 0.1_497, 0.1_513, 0.1_505, 0.1_442, 0.1_453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] ='google/ddpm-ema-bedroom-256' __UpperCamelCase : Any =UNetaDModel.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : int =DDIMScheduler.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : Dict =DDIMPipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) ddpm.to(lowerCamelCase__ ) ddpm.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple =torch.manual_seed(0 ) __UpperCamelCase : Union[str, Any] =ddpm(generator=lowerCamelCase__ , output_type='numpy' ).images __UpperCamelCase : Tuple =image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __UpperCamelCase : Optional[Any] =np.array([0.0_060, 0.0_201, 0.0_344, 0.0_024, 0.0_018, 0.0_002, 0.0_022, 0.0_000, 0.0_069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
71
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A_ :Union[str, Any] = logging.get_logger(__name__) A_ :List[Any] = { '''asapp/sew-d-tiny-100k''': '''https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json''', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[int] ="""sew-d""" def __init__( self , lowerCamelCase__=32 , lowerCamelCase__=768 , lowerCamelCase__=12 , lowerCamelCase__=12 , lowerCamelCase__=3072 , lowerCamelCase__=2 , lowerCamelCase__=512 , lowerCamelCase__=256 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=("p2c", "c2p") , lowerCamelCase__="layer_norm" , lowerCamelCase__="gelu_python" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0 , lowerCamelCase__=0.1 , lowerCamelCase__=0.02 , lowerCamelCase__=1E-7 , lowerCamelCase__=1E-5 , lowerCamelCase__="group" , lowerCamelCase__="gelu" , lowerCamelCase__=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowerCamelCase__=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowerCamelCase__=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowerCamelCase__=False , lowerCamelCase__=128 , lowerCamelCase__=16 , lowerCamelCase__=True , lowerCamelCase__=0.05 , lowerCamelCase__=10 , lowerCamelCase__=2 , lowerCamelCase__=0.0 , lowerCamelCase__=10 , lowerCamelCase__=0 , lowerCamelCase__="mean" , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=256 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ , pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ ) __UpperCamelCase : Dict =hidden_size __UpperCamelCase : List[Any] =feat_extract_norm __UpperCamelCase : Union[str, Any] =feat_extract_activation __UpperCamelCase : Union[str, Any] =list(lowerCamelCase__ ) __UpperCamelCase : str =list(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =list(lowerCamelCase__ ) __UpperCamelCase : Optional[int] =conv_bias __UpperCamelCase : List[Any] =num_conv_pos_embeddings __UpperCamelCase : Optional[Any] =num_conv_pos_embedding_groups __UpperCamelCase : List[str] =len(self.conv_dim ) __UpperCamelCase : Union[str, Any] =num_hidden_layers __UpperCamelCase : Optional[Any] =intermediate_size __UpperCamelCase : Tuple =squeeze_factor __UpperCamelCase : Union[str, Any] =max_position_embeddings __UpperCamelCase : List[str] =position_buckets __UpperCamelCase : Any =share_att_key __UpperCamelCase : List[Any] =relative_attention __UpperCamelCase : str =norm_rel_ebd __UpperCamelCase : Optional[Any] =list(lowerCamelCase__ ) __UpperCamelCase : List[Any] =hidden_act __UpperCamelCase : int =num_attention_heads __UpperCamelCase : List[str] =hidden_dropout __UpperCamelCase : str =attention_dropout __UpperCamelCase : str =activation_dropout __UpperCamelCase : Dict =feat_proj_dropout __UpperCamelCase : Optional[Any] =final_dropout __UpperCamelCase : Any =layer_norm_eps __UpperCamelCase : List[Any] =feature_layer_norm_eps __UpperCamelCase : Tuple =initializer_range __UpperCamelCase : Any =vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect.' 'It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,' f'but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)' f'= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCamelCase : Union[str, Any] =apply_spec_augment __UpperCamelCase : str =mask_time_prob __UpperCamelCase : int =mask_time_length __UpperCamelCase : List[str] =mask_time_min_masks __UpperCamelCase : Dict =mask_feature_prob __UpperCamelCase : Tuple =mask_feature_length __UpperCamelCase : Tuple =mask_feature_min_masks # ctc loss __UpperCamelCase : Optional[int] =ctc_loss_reduction __UpperCamelCase : Union[str, Any] =ctc_zero_infinity # sequence classification __UpperCamelCase : Dict =use_weighted_layer_sum __UpperCamelCase : Union[str, Any] =classifier_proj_size @property def __lowercase ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
71
from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[int] ="""new-model""" if is_tf_available(): class __A ( a ): """simple docstring""" UpperCamelCase__ : List[str] =NewModelConfig @require_tf class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ='bert-base-cased' __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] ='bert-base-cased' __UpperCamelCase : Optional[int] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =TFAutoModelForPreTraining.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : str =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : str =TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Optional[int] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Any =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Any =TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : str =TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =TFAutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow @require_tensorflow_probability def __lowercase ( self ): """simple docstring""" for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModelForTableQuestionAnswering.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[str] =TFAutoModelForTableQuestionAnswering.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =copy.deepcopy(model.config ) __UpperCamelCase : Optional[Any] =['FunnelBaseModel'] __UpperCamelCase : Tuple =TFAutoModel.from_config(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : List[Any] =TFAutoModel.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" try: AutoConfig.register('new-model' , lowerCamelCase__ ) __UpperCamelCase : int =[ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowerCamelCase__ ): auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCamelCase__ ): auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) # Now that the config is registered, it can be used as any other config with the auto-API __UpperCamelCase : List[str] =BertModelTester(self ).get_config() __UpperCamelCase : Optional[Any] =NewModelConfig(**tiny_config.to_dict() ) __UpperCamelCase : Dict =auto_class.from_config(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =auto_class.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex( lowerCamelCase__ , 'bert-base is not a local folder and is not a valid model identifier' ): __UpperCamelCase : Dict =TFAutoModel.from_pretrained('bert-base' ) def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex( lowerCamelCase__ , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): __UpperCamelCase : Union[str, Any] =TFAutoModel.from_pretrained(lowerCamelCase__ , revision='aaaaaa' ) def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex( lowerCamelCase__ , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): __UpperCamelCase : List[str] =TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex(lowerCamelCase__ , 'Use `from_pt=True` to load this model' ): __UpperCamelCase : List[Any] =TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: __UpperCamelCase : Dict =TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint __UpperCamelCase : Dict =TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: __UpperCamelCase : Union[str, Any] =TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
71
1
import argparse import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A_ :Optional[int] = 16 A_ :List[str] = 32 def A ( a_ ,a_ = 16 ) -> List[str]: __UpperCamelCase : List[Any] =AutoTokenizer.from_pretrained('bert-base-cased' ) __UpperCamelCase : Tuple =load_dataset('glue' ,'mrpc' ) def tokenize_function(a_ ): # max_length=None => use the model max length (it's actually the default) __UpperCamelCase : Tuple =tokenizer(examples['sentence1'] ,examples['sentence2'] ,truncation=a_ ,max_length=a_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __UpperCamelCase : Any =datasets.map( a_ ,batched=a_ ,remove_columns=['idx', 'sentence1', 'sentence2'] ,) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __UpperCamelCase : List[Any] =tokenized_datasets.rename_column('label' ,'labels' ) def collate_fn(a_ ): # On TPU it's best to pad everything to the same length or training will be very slow. __UpperCamelCase : Tuple =128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __UpperCamelCase : List[Any] =16 elif accelerator.mixed_precision != "no": __UpperCamelCase : Optional[Any] =8 else: __UpperCamelCase : Any =None return tokenizer.pad( a_ ,padding='longest' ,max_length=a_ ,pad_to_multiple_of=a_ ,return_tensors='pt' ,) # Instantiate dataloaders. __UpperCamelCase : List[str] =DataLoader( tokenized_datasets['train'] ,shuffle=a_ ,collate_fn=a_ ,batch_size=a_ ,drop_last=a_ ) __UpperCamelCase : Optional[Any] =DataLoader( tokenized_datasets['validation'] ,shuffle=a_ ,collate_fn=a_ ,batch_size=a_ ,drop_last=(accelerator.mixed_precision == 'fp8') ,) return train_dataloader, eval_dataloader def A ( a_ ,a_ ) -> List[Any]: # Initialize accelerator __UpperCamelCase : str =Accelerator(cpu=args.cpu ,mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __UpperCamelCase : Tuple =config['lr'] __UpperCamelCase : Tuple =int(config['num_epochs'] ) __UpperCamelCase : Tuple =int(config['seed'] ) __UpperCamelCase : Tuple =int(config['batch_size'] ) __UpperCamelCase : Dict =evaluate.load('glue' ,'mrpc' ) # If the batch size is too big we use gradient accumulation __UpperCamelCase : int =1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: __UpperCamelCase : Optional[Any] =batch_size // MAX_GPU_BATCH_SIZE __UpperCamelCase : int =MAX_GPU_BATCH_SIZE set_seed(a_ ) __UpperCamelCase , __UpperCamelCase : List[Any] =get_dataloaders(a_ ,a_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __UpperCamelCase : Union[str, Any] =AutoModelForSequenceClassification.from_pretrained('bert-base-cased' ,return_dict=a_ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __UpperCamelCase : Union[str, Any] =model.to(accelerator.device ) # Instantiate optimizer __UpperCamelCase : Optional[int] =AdamW(params=model.parameters() ,lr=a_ ) # Instantiate scheduler __UpperCamelCase : Dict =get_linear_schedule_with_warmup( optimizer=a_ ,num_warmup_steps=100 ,num_training_steps=(len(a_ ) * num_epochs) // gradient_accumulation_steps ,) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase : Dict =accelerator.prepare( a_ ,a_ ,a_ ,a_ ,a_ ) # Now we train the model for epoch in range(a_ ): model.train() for step, batch in enumerate(a_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) __UpperCamelCase : Union[str, Any] =model(**a_ ) __UpperCamelCase : Any =outputs.loss __UpperCamelCase : List[Any] =loss / gradient_accumulation_steps accelerator.backward(a_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(a_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __UpperCamelCase : List[Any] =model(**a_ ) __UpperCamelCase : Optional[int] =outputs.logits.argmax(dim=-1 ) __UpperCamelCase , __UpperCamelCase : Tuple =accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=a_ ,references=a_ ,) __UpperCamelCase : int =metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}:' ,a_ ) def A ( ) -> Union[str, Any]: __UpperCamelCase : Union[str, Any] =argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' ,type=a_ ,default=a_ ,choices=['no', 'fp16', 'bf16', 'fp8'] ,help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' ,) parser.add_argument('--cpu' ,action='store_true' ,help='If passed, will train on the CPU.' ) __UpperCamelCase : List[str] =parser.parse_args() __UpperCamelCase : List[Any] ={'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(a_ ,a_ ) if __name__ == "__main__": main()
71
import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() A_ :List[str] = [ '''word_embeddings_layernorm.weight''', '''word_embeddings_layernorm.bias''', '''input_layernorm.weight''', '''input_layernorm.bias''', '''post_attention_layernorm.weight''', '''post_attention_layernorm.bias''', '''self_attention.dense.bias''', '''mlp.dense_4h_to_h.bias''', '''ln_f.weight''', '''ln_f.bias''', ] A_ :Optional[Any] = [ '''mlp.dense_4h_to_h.weight''', '''self_attention.dense.weight''', ] def A ( a_ ,a_ ) -> str: __UpperCamelCase : Any ={ 'word_embeddings.weight': 'word_embeddings.weight', 'word_embeddings.norm.weight': 'word_embeddings_layernorm.weight', 'word_embeddings.norm.bias': 'word_embeddings_layernorm.bias', 'weight': 'ln_f.weight', 'bias': 'ln_f.bias', } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks __UpperCamelCase : Tuple =int(re.match(r'.*layer_(\d*).*' ,a_ )[1] ) layer_number -= 3 return F'h.{layer_number}.' + key def A ( a_ ) -> Any: if dtype == torch.bool: return 1 / 8 __UpperCamelCase : Dict =re.search(r'[^\d](\d+)$' ,str(a_ ) ) if bit_search is None: raise ValueError(F'`dtype` is not a valid dtype: {dtype}.' ) __UpperCamelCase : Tuple =int(bit_search.groups()[0] ) return bit_size // 8 def A ( a_ ,a_ ,a_ ,a_ ,a_ ) -> Dict: # Construct model if bloom_config_file == "": __UpperCamelCase : List[Any] =BloomConfig() else: __UpperCamelCase : List[str] =BloomConfig.from_json_file(a_ ) if shard_model: __UpperCamelCase : int =os.listdir(a_ ) __UpperCamelCase : Union[str, Any] =sorted(filter(lambda a_ : s.startswith('layer' ) and "model_00" in s ,a_ ) ) __UpperCamelCase : Optional[Any] ={'weight_map': {}, 'metadata': {}} __UpperCamelCase : Dict =0 __UpperCamelCase : int =None __UpperCamelCase : Any =BloomConfig() for j, file in enumerate(a_ ): print('Processing file: {}'.format(a_ ) ) __UpperCamelCase : Optional[int] =None for i in range(a_ ): # load all TP files __UpperCamelCase : Dict =file.replace('model_00' ,F'model_0{i}' ) __UpperCamelCase : Optional[Any] =torch.load(os.path.join(a_ ,a_ ) ,map_location='cpu' ) # Rename keys in the transformers names __UpperCamelCase : int =list(temp.keys() ) for key in keys: __UpperCamelCase : Dict =temp.pop(a_ ) if tensors is None: __UpperCamelCase : Any =temp else: for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel __UpperCamelCase : List[Any] =1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks __UpperCamelCase : Any =torch.cat([tensors[key], temp[key]] ,dim=a_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): __UpperCamelCase : Optional[Any] =tensors[key] / pretraining_tp torch.save( a_ ,os.path.join( a_ ,'pytorch_model_{}-of-{}.bin'.format(str(j + 1 ).zfill(5 ) ,str(len(a_ ) ).zfill(5 ) ) ,) ,) for key in tensors.keys(): __UpperCamelCase : Union[str, Any] =tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: __UpperCamelCase : int ='pytorch_model_{}-of-{}.bin'.format( str(j + 1 ).zfill(5 ) ,str(len(a_ ) ).zfill(5 ) ) __UpperCamelCase : Union[str, Any] =BloomConfig() __UpperCamelCase : Tuple =pytorch_dump_folder_path + '/' + CONFIG_NAME __UpperCamelCase : Optional[int] =total_size with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) with open(os.path.join(a_ ,WEIGHTS_NAME + '.index.json' ) ,'w' ,encoding='utf-8' ) as f: __UpperCamelCase : List[Any] =json.dumps(a_ ,indent=2 ,sort_keys=a_ ) + '\n' f.write(a_ ) else: __UpperCamelCase : List[Any] =BloomModel(a_ ) __UpperCamelCase : Optional[Any] =os.listdir(a_ ) __UpperCamelCase : Dict =sorted(filter(lambda a_ : s.startswith('layer' ) and "model_00" in s ,a_ ) ) __UpperCamelCase : Any =None for i, file in enumerate(a_ ): __UpperCamelCase : Union[str, Any] =None for i in range(a_ ): # load all TP files __UpperCamelCase : Optional[Any] =file.replace('model_00' ,F'model_0{i}' ) __UpperCamelCase : str =torch.load(os.path.join(a_ ,a_ ) ,map_location='cpu' ) # Rename keys in the transformers names __UpperCamelCase : List[str] =list(temp.keys() ) for key in keys: __UpperCamelCase : Union[str, Any] =temp.pop(a_ ) if tensors is None: __UpperCamelCase : Optional[Any] =temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel __UpperCamelCase : Optional[int] =1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks __UpperCamelCase : int =torch.cat([tensors[key], temp[key]] ,dim=a_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): __UpperCamelCase : Dict =tensors[key] / pretraining_tp __UpperCamelCase : str =model.load_state_dict(a_ ,strict=a_ ) assert not other_keys.unexpected_keys, F'The keys {other_keys.unexpected_keys} are unexpected' if missing_keys is None: __UpperCamelCase : str =set(other_keys.missing_keys ) else: __UpperCamelCase : int =missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F'The keys {missing_keys} are missing' # Save pytorch-model os.makedirs(a_ ,exist_ok=a_ ) __UpperCamelCase : Optional[int] =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase : Dict =pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}' ) if config.torch_dtype is not None: __UpperCamelCase : List[str] =model.to(config.torch_dtype ) torch.save(model.state_dict() ,a_ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A_ :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bloom_checkpoint_path''', default=None, type=str, required=True, help='''Path to the Megatron-LM checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--bloom_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--shard_model''', action='''store_true''', help='''An optional setting to shard the output model \nThis enables sharding the converted checkpoint''', ) parser.add_argument( '''--pretraining_tp''', default=4, type=int, help='''Pretraining TP rank that has been used when training the model in Megatron-LM \n''', ) A_ :str = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
71
1
import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict A_ :Tuple = namedtuple( '''_TestCommandArgs''', [ '''dataset''', '''name''', '''cache_dir''', '''data_dir''', '''all_configs''', '''save_infos''', '''ignore_verifications''', '''force_redownload''', '''clear_cache''', ], defaults=[None, None, None, False, False, False, False, False], ) def A ( a_ ,a_ ) -> Dict: return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def A ( a_ ) -> Dict: __UpperCamelCase : Optional[int] =_TestCommandArgs(dataset=a_ ,all_configs=a_ ,save_infos=a_ ) __UpperCamelCase : Any =TestCommand(*a_ ) test_command.run() __UpperCamelCase : List[str] =os.path.join(a_ ,'README.md' ) assert os.path.exists(a_ ) __UpperCamelCase : Any =DatasetInfosDict.from_directory(a_ ) __UpperCamelCase : int =DatasetInfosDict( { 'default': DatasetInfo( features=Features( { 'tokens': Sequence(Value('string' ) ), 'ner_tags': Sequence( ClassLabel(names=['O', 'B-PER', 'I-PER', 'B-ORG', 'I-ORG', 'B-LOC', 'I-LOC'] ) ), 'langs': Sequence(Value('string' ) ), 'spans': Sequence(Value('string' ) ), } ) ,splits=[ { 'name': 'train', 'num_bytes': 2_351_563, 'num_examples': 10_000, }, { 'name': 'validation', 'num_bytes': 238_418, 'num_examples': 1_000, }, ] ,download_size=3_940_680 ,dataset_size=2_589_981 ,) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: __UpperCamelCase , __UpperCamelCase : Optional[Any] =getattr(dataset_infos['default'] ,a_ ), getattr(expected_dataset_infos['default'] ,a_ ) if key == "num_bytes": assert is_apercent_close(a_ ,a_ ) elif key == "splits": assert list(a_ ) == list(a_ ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes ,expected[split].num_bytes ) else: result == expected
71
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class __A : """simple docstring""" UpperCamelCase__ : int =XGLMConfig UpperCamelCase__ : Optional[Any] ={} UpperCamelCase__ : List[str] ="""gelu""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=14 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=99 , lowerCamelCase__=32 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=37 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=512 , lowerCamelCase__=0.02 , ): """simple docstring""" __UpperCamelCase : Tuple =parent __UpperCamelCase : List[str] =batch_size __UpperCamelCase : str =seq_length __UpperCamelCase : Dict =is_training __UpperCamelCase : Tuple =use_input_mask __UpperCamelCase : List[Any] =use_labels __UpperCamelCase : Any =vocab_size __UpperCamelCase : List[Any] =d_model __UpperCamelCase : Optional[int] =num_hidden_layers __UpperCamelCase : List[str] =num_attention_heads __UpperCamelCase : Optional[int] =ffn_dim __UpperCamelCase : str =activation_function __UpperCamelCase : Any =activation_dropout __UpperCamelCase : Optional[int] =attention_dropout __UpperCamelCase : Optional[int] =max_position_embeddings __UpperCamelCase : Any =initializer_range __UpperCamelCase : Dict =None __UpperCamelCase : Optional[int] =0 __UpperCamelCase : Optional[Any] =2 __UpperCamelCase : str =1 def __lowercase ( self ): """simple docstring""" return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) __UpperCamelCase : Union[str, Any] =None if self.use_input_mask: __UpperCamelCase : Dict =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase : Any =self.get_config() __UpperCamelCase : Optional[Any] =floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def __lowercase ( self ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=lowerCamelCase__ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=lowerCamelCase__ , ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) : int =config_and_inputs __UpperCamelCase : Optional[Any] ={ 'input_ids': input_ids, 'head_mask': head_mask, } return config, inputs_dict @require_tf class __A ( a , a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Union[str, Any] =(TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () UpperCamelCase__ : str =(TFXGLMForCausalLM,) if is_tf_available() else () UpperCamelCase__ : Optional[Any] =( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) UpperCamelCase__ : Tuple =False UpperCamelCase__ : Tuple =False UpperCamelCase__ : Optional[Any] =False def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =TFXGLMModelTester(self ) __UpperCamelCase : Dict =ConfigTester(self , config_class=lowerCamelCase__ , n_embd=37 ) def __lowercase ( self ): """simple docstring""" self.config_tester.run_common_tests() @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Optional[Any] =TFXGLMModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def __lowercase ( self ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self , lowerCamelCase__=True ): """simple docstring""" __UpperCamelCase : int =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : List[str] =tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off __UpperCamelCase : str =[2, 268, 9865, 67, 11, 1988, 57252, 9865, 5, 984, 67, 1988, 213838, 1658, 53, 70446, 33, 6657, 278, 1581] # fmt: on __UpperCamelCase : Optional[Any] =model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Union[str, Any] =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) __UpperCamelCase : str =tokenizer('Today is a nice day and' , return_tensors='tf' ) __UpperCamelCase : Union[str, Any] =tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): __UpperCamelCase : Any =model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ , seed=[7, 0] ) __UpperCamelCase : Tuple =tokenizer.decode(output_ids[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : List[Any] =( 'Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due' ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Optional[Any] =XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Optional[Any] ='left' # use different length sentences to test batching __UpperCamelCase : Optional[int] =[ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When', 'Hello, my dog is a little', ] __UpperCamelCase : List[Any] =tokenizer(lowerCamelCase__ , return_tensors='tf' , padding=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =inputs['input_ids'] __UpperCamelCase : Dict =model.generate(input_ids=lowerCamelCase__ , attention_mask=inputs['attention_mask'] , max_new_tokens=12 ) __UpperCamelCase : List[Any] =tokenizer(sentences[0] , return_tensors='tf' ).input_ids __UpperCamelCase : Dict =model.generate(input_ids=lowerCamelCase__ , max_new_tokens=12 ) __UpperCamelCase : Any =tokenizer(sentences[1] , return_tensors='tf' ).input_ids __UpperCamelCase : Optional[Any] =model.generate(input_ids=lowerCamelCase__ , max_new_tokens=12 ) __UpperCamelCase : Optional[int] =tokenizer.batch_decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : int =tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : Any =[ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ' 'a single', 'Hello, my dog is a little bit of a shy one, but he is very friendly', ] self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , [non_padded_sentence, padded_sentence] )
71
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_lxmert import LxmertTokenizer A_ :Dict = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ :Dict = { '''vocab_file''': { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/vocab.txt''', }, '''tokenizer_file''': { '''unc-nlp/lxmert-base-uncased''': ( '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ :str = { '''unc-nlp/lxmert-base-uncased''': 512, } A_ :Dict = { '''unc-nlp/lxmert-base-uncased''': {'''do_lower_case''': True}, } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[Any] =VOCAB_FILES_NAMES UpperCamelCase__ : List[str] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : Union[str, Any] =PRETRAINED_INIT_CONFIGURATION UpperCamelCase__ : int =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : str =LxmertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): """simple docstring""" super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) __UpperCamelCase : Dict =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('strip_accents' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowerCamelCase__ ) != tokenize_chinese_chars ): __UpperCamelCase : str =getattr(lowerCamelCase__ , normalizer_state.pop('type' ) ) __UpperCamelCase : Any =do_lower_case __UpperCamelCase : Dict =strip_accents __UpperCamelCase : List[str] =tokenize_chinese_chars __UpperCamelCase : Union[str, Any] =normalizer_class(**lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =do_lower_case def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=None ): """simple docstring""" __UpperCamelCase : List[str] =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : str =[self.sep_token_id] __UpperCamelCase : Union[str, Any] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : int =self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
71
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A ( a_ ,a_ ) -> Optional[Any]: # Load checkpoint __UpperCamelCase : int =torch.load(a_ ,map_location='cpu' ) __UpperCamelCase : List[Any] =chkpt['model'] # We have the base model one level deeper than the original XLM repository __UpperCamelCase : str ={} for k, v in state_dict.items(): if "pred_layer" in k: __UpperCamelCase : Optional[Any] =v else: __UpperCamelCase : Optional[Any] =v __UpperCamelCase : List[Any] =chkpt['params'] __UpperCamelCase : str ={n: v for n, v in config.items() if not isinstance(a_ ,(torch.FloatTensor, numpy.ndarray) )} __UpperCamelCase : str =chkpt['dico_word2id'] __UpperCamelCase : Dict ={s + '</w>' if s.find('@@' ) == -1 and i > 13 else s.replace('@@' ,'' ): i for s, i in vocab.items()} # Save pytorch-model __UpperCamelCase : List[Any] =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase : Tuple =pytorch_dump_folder_path + '/' + CONFIG_NAME __UpperCamelCase : Any =pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['vocab_file'] print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(a_ ,a_ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(a_ ,indent=2 ) + '\n' ) print(F'Save vocab file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(a_ ,indent=2 ) + '\n' ) if __name__ == "__main__": A_ :str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xlm_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) A_ :List[Any] = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
71
1
import csv import tweepy # Twitter API credentials A_ :int = '''''' A_ :str = '''''' A_ :Dict = '''''' A_ :Optional[int] = '''''' def A ( a_ ) -> None: # authorize twitter, initialize tweepy __UpperCamelCase : str =tweepy.OAuthHandler(a_ ,a_ ) auth.set_access_token(a_ ,a_ ) __UpperCamelCase : Optional[int] =tweepy.API(a_ ) # initialize a list to hold all the tweepy Tweets __UpperCamelCase : Optional[Any] =[] # make initial request for most recent tweets (200 is the maximum allowed count) __UpperCamelCase : str =api.user_timeline(screen_name=a_ ,count=200 ) # save most recent tweets alltweets.extend(a_ ) # save the id of the oldest tweet less one __UpperCamelCase : Any =alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(a_ ) > 0: print(F'getting tweets before {oldest}' ) # all subsequent requests use the max_id param to prevent duplicates __UpperCamelCase : Optional[int] =api.user_timeline( screen_name=a_ ,count=200 ,max_id=a_ ) # save most recent tweets alltweets.extend(a_ ) # update the id of the oldest tweet less one __UpperCamelCase : Optional[Any] =alltweets[-1].id - 1 print(F'...{len(a_ )} tweets downloaded so far' ) # transform the tweepy tweets into a 2D array that will populate the csv __UpperCamelCase : Dict =[[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(F'new_{screen_name}_tweets.csv' ,'w' ) as f: __UpperCamelCase : Optional[int] =csv.writer(a_ ) writer.writerow(['id', 'created_at', 'text'] ) writer.writerows(a_ ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets('''FirePing32''')
71
import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __A ( a ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =tempfile.mkdtemp() __UpperCamelCase : Optional[int] =8 # DPR tok __UpperCamelCase : str =[ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase : Optional[Any] =os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase : Optional[int] =[ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase : str =dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) __UpperCamelCase : Optional[int] =['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase : Any ={'unk_token': '<unk>'} __UpperCamelCase : Any =os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Any =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(lowerCamelCase__ ) ) def __lowercase ( self ): """simple docstring""" return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) @require_tokenizers def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =os.path.join(self.tmpdirname , 'rag_tokenizer' ) __UpperCamelCase : Dict =RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) __UpperCamelCase : List[Any] =RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(lowerCamelCase__ ) rag_tokenizer.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : int =RagTokenizer.from_pretrained(lowerCamelCase__ , config=lowerCamelCase__ ) self.assertIsInstance(new_rag_tokenizer.question_encoder , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =RagTokenizer.from_pretrained('facebook/rag-token-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : int =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =RagTokenizer.from_pretrained('facebook/rag-sequence-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : Any =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
71
1
import re from filelock import FileLock try: import nltk A_ :Any = True except (ImportError, ModuleNotFoundError): A_ :Union[str, Any] = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def A ( a_ ) -> str: re.sub('<n>' ,'' ,a_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(a_ ) )
71
A_ :Optional[int] = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' A_ :Union[str, Any] = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] A_ :Optional[Any] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
71
1
from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging A_ :Tuple = logging.get_logger(__name__) class __A ( a ): """simple docstring""" UpperCamelCase__ : List[Any] =["""pixel_values"""] def __init__( self , lowerCamelCase__ = True , lowerCamelCase__ = 32 , lowerCamelCase__=PILImageResampling.BILINEAR , lowerCamelCase__ = True , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Dict =do_resize __UpperCamelCase : int =do_rescale __UpperCamelCase : Optional[Any] =size_divisor __UpperCamelCase : str =resample super().__init__(**lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Union[str, Any] =get_image_size(lowerCamelCase__ ) # Rounds the height and width down to the closest multiple of size_divisor __UpperCamelCase : Union[str, Any] =height // size_divisor * size_divisor __UpperCamelCase : Union[str, Any] =width // size_divisor * size_divisor __UpperCamelCase : Tuple =resize(lowerCamelCase__ , (new_h, new_w) , resample=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) return image def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ ): """simple docstring""" return rescale(image=lowerCamelCase__ , scale=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__=None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = ChannelDimension.FIRST , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[int] =do_resize if do_resize is not None else self.do_resize __UpperCamelCase : List[Any] =do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase : Optional[Any] =size_divisor if size_divisor is not None else self.size_divisor __UpperCamelCase : Union[str, Any] =resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('size_divisor is required for resizing' ) __UpperCamelCase : List[str] =make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError('Invalid image(s)' ) # All transformations expect numpy arrays. __UpperCamelCase : str =[to_numpy_array(lowerCamelCase__ ) for img in images] if do_resize: __UpperCamelCase : Optional[int] =[self.resize(lowerCamelCase__ , size_divisor=lowerCamelCase__ , resample=lowerCamelCase__ ) for image in images] if do_rescale: __UpperCamelCase : List[Any] =[self.rescale(lowerCamelCase__ , scale=1 / 255 ) for image in images] __UpperCamelCase : Tuple =[to_channel_dimension_format(lowerCamelCase__ , lowerCamelCase__ ) for image in images] __UpperCamelCase : Dict ={'pixel_values': images} return BatchFeature(data=lowerCamelCase__ , tensor_type=lowerCamelCase__ )
71
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration A_ :Optional[Any] = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def A ( a_ ) -> List[Any]: __UpperCamelCase : Any =['layers', 'blocks'] for k in ignore_keys: state_dict.pop(a_ ,a_ ) A_ :int = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def A ( a_ ) -> Union[str, Any]: __UpperCamelCase : str =list(s_dict.keys() ) for key in keys: __UpperCamelCase : str =key for k, v in WHISPER_MAPPING.items(): if k in key: __UpperCamelCase : Optional[Any] =new_key.replace(a_ ,a_ ) print(F'{key} -> {new_key}' ) __UpperCamelCase : Dict =s_dict.pop(a_ ) return s_dict def A ( a_ ) -> Optional[Any]: __UpperCamelCase , __UpperCamelCase : Tuple =emb.weight.shape __UpperCamelCase : Tuple =nn.Linear(a_ ,a_ ,bias=a_ ) __UpperCamelCase : List[Any] =emb.weight.data return lin_layer def A ( a_ ,a_ ) -> bytes: os.makedirs(a_ ,exist_ok=a_ ) __UpperCamelCase : Optional[int] =os.path.basename(a_ ) __UpperCamelCase : Union[str, Any] =url.split('/' )[-2] __UpperCamelCase : Union[str, Any] =os.path.join(a_ ,a_ ) if os.path.exists(a_ ) and not os.path.isfile(a_ ): raise RuntimeError(F'{download_target} exists and is not a regular file' ) if os.path.isfile(a_ ): __UpperCamelCase : str =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(F'{download_target} exists, but the SHA256 checksum does not match; re-downloading the file' ) with urllib.request.urlopen(a_ ) as source, open(a_ ,'wb' ) as output: with tqdm( total=int(source.info().get('Content-Length' ) ) ,ncols=80 ,unit='iB' ,unit_scale=a_ ,unit_divisor=1_024 ) as loop: while True: __UpperCamelCase : Optional[Any] =source.read(8_192 ) if not buffer: break output.write(a_ ) loop.update(len(a_ ) ) __UpperCamelCase : List[Any] =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() != expected_shaaaa: raise RuntimeError( 'Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.' ) return model_bytes def A ( a_ ,a_ ) -> Optional[Any]: if ".pt" not in checkpoint_path: __UpperCamelCase : int =_download(_MODELS[checkpoint_path] ) else: __UpperCamelCase : List[str] =torch.load(a_ ,map_location='cpu' ) __UpperCamelCase : Union[str, Any] =original_checkpoint['dims'] __UpperCamelCase : List[Any] =original_checkpoint['model_state_dict'] __UpperCamelCase : Dict =state_dict['decoder.token_embedding.weight'] remove_ignore_keys_(a_ ) rename_keys(a_ ) __UpperCamelCase : List[str] =True __UpperCamelCase : str =state_dict['decoder.layers.0.fc1.weight'].shape[0] __UpperCamelCase : Optional[int] =WhisperConfig( vocab_size=dimensions['n_vocab'] ,encoder_ffn_dim=a_ ,decoder_ffn_dim=a_ ,num_mel_bins=dimensions['n_mels'] ,d_model=dimensions['n_audio_state'] ,max_target_positions=dimensions['n_text_ctx'] ,encoder_layers=dimensions['n_audio_layer'] ,encoder_attention_heads=dimensions['n_audio_head'] ,decoder_layers=dimensions['n_text_layer'] ,decoder_attention_heads=dimensions['n_text_state'] ,max_source_positions=dimensions['n_audio_ctx'] ,) __UpperCamelCase : List[str] =WhisperForConditionalGeneration(a_ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =model.model.load_state_dict(a_ ,strict=a_ ) if len(a_ ) > 0 and not set(a_ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( 'Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,' F' but all the following weights are missing {missing}' ) if tie_embeds: __UpperCamelCase : Optional[int] =make_linear_from_emb(model.model.decoder.embed_tokens ) else: __UpperCamelCase : List[str] =proj_out_weights model.save_pretrained(a_ ) if __name__ == "__main__": A_ :List[Any] = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') A_ :List[Any] = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
71
1
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
71
import os from datetime import datetime as dt from github import Github A_ :str = [ '''good first issue''', '''feature request''', '''wip''', ] def A ( ) -> Any: __UpperCamelCase : Any =Github(os.environ['GITHUB_TOKEN'] ) __UpperCamelCase : Union[str, Any] =g.get_repo('huggingface/accelerate' ) __UpperCamelCase : Tuple =repo.get_issues(state='open' ) for issue in open_issues: __UpperCamelCase : List[Any] =sorted([comment for comment in issue.get_comments()] ,key=lambda a_ : i.created_at ,reverse=a_ ) __UpperCamelCase : str =comments[0] if len(a_ ) > 0 else None __UpperCamelCase : Any =dt.utcnow() __UpperCamelCase : List[str] =(current_time - issue.updated_at).days __UpperCamelCase : Union[str, Any] =(current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state='closed' ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
71
1
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def A ( a_ ) -> int: __UpperCamelCase : Optional[Any] =filter(lambda a_ : p.requires_grad ,model.parameters() ) __UpperCamelCase : Tuple =sum([np.prod(p.size() ) for p in model_parameters] ) return params A_ :Tuple = logging.getLogger(__name__) def A ( a_ ,a_ ) -> str: if metric == "rouge2": __UpperCamelCase : str ='{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __UpperCamelCase : Optional[int] ='{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __UpperCamelCase : Optional[Any] ='{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F'seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this' ' function.' ) __UpperCamelCase : Optional[Any] =ModelCheckpoint( dirpath=a_ ,filename=a_ ,monitor=F'val_{metric}' ,mode='max' ,save_top_k=3 ,every_n_epochs=1 ,) return checkpoint_callback def A ( a_ ,a_ ) -> List[Any]: return EarlyStopping( monitor=F'val_{metric}' ,mode='min' if 'loss' in metric else 'max' ,patience=a_ ,verbose=a_ ,) class __A ( pl.Callback ): """simple docstring""" def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] ={f'lr_group_{i}': param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCamelCase__ ) @rank_zero_only def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=True ): """simple docstring""" logger.info(f'***** {type_path} results at step {trainer.global_step:05d} *****' ) __UpperCamelCase : Tuple =trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __UpperCamelCase : Optional[Any] =Path(pl_module.hparams.output_dir ) if type_path == "test": __UpperCamelCase : int =od / 'test_results.txt' __UpperCamelCase : Union[str, Any] =od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __UpperCamelCase : Optional[Any] =od / f'{type_path}_results/{trainer.global_step:05d}.txt' __UpperCamelCase : Any =od / f'{type_path}_generations/{trainer.global_step:05d}.txt' results_file.parent.mkdir(exist_ok=lowerCamelCase__ ) generations_file.parent.mkdir(exist_ok=lowerCamelCase__ ) with open(lowerCamelCase__ , 'a+' ) as writer: for key in sorted(lowerCamelCase__ ): if key in ["log", "progress_bar", "preds"]: continue __UpperCamelCase : Optional[Any] =metrics[key] if isinstance(lowerCamelCase__ , torch.Tensor ): __UpperCamelCase : Dict =val.item() __UpperCamelCase : int =f'{key}: {val:.6f}\n' writer.write(lowerCamelCase__ ) if not save_generations: return if "preds" in metrics: __UpperCamelCase : List[str] ='\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCamelCase__ ) @rank_zero_only def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" try: __UpperCamelCase : Optional[Any] =pl_module.model.model.num_parameters() except AttributeError: __UpperCamelCase : Optional[Any] =pl_module.model.num_parameters() __UpperCamelCase : Optional[Any] =count_trainable_parameters(lowerCamelCase__ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1E6, 'grad_mp': n_trainable_pars / 1E6} ) @rank_zero_only def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCamelCase__ , lowerCamelCase__ , 'test' ) @rank_zero_only def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
71
import re def A ( a_ ) -> bool: __UpperCamelCase : Any =re.compile( r'^(?:0|94|\+94|0{2}94)' r'7(0|1|2|4|5|6|7|8)' r'(-| |)' r'\d{7}$' ) return bool(re.search(a_ ,a_ ) ) if __name__ == "__main__": A_ :List[str] = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
71
1
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration A_ :Optional[Any] = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def A ( a_ ) -> List[Any]: __UpperCamelCase : Any =['layers', 'blocks'] for k in ignore_keys: state_dict.pop(a_ ,a_ ) A_ :int = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def A ( a_ ) -> Union[str, Any]: __UpperCamelCase : str =list(s_dict.keys() ) for key in keys: __UpperCamelCase : str =key for k, v in WHISPER_MAPPING.items(): if k in key: __UpperCamelCase : Optional[Any] =new_key.replace(a_ ,a_ ) print(F'{key} -> {new_key}' ) __UpperCamelCase : Dict =s_dict.pop(a_ ) return s_dict def A ( a_ ) -> Optional[Any]: __UpperCamelCase , __UpperCamelCase : Tuple =emb.weight.shape __UpperCamelCase : Tuple =nn.Linear(a_ ,a_ ,bias=a_ ) __UpperCamelCase : List[Any] =emb.weight.data return lin_layer def A ( a_ ,a_ ) -> bytes: os.makedirs(a_ ,exist_ok=a_ ) __UpperCamelCase : Optional[int] =os.path.basename(a_ ) __UpperCamelCase : Union[str, Any] =url.split('/' )[-2] __UpperCamelCase : Union[str, Any] =os.path.join(a_ ,a_ ) if os.path.exists(a_ ) and not os.path.isfile(a_ ): raise RuntimeError(F'{download_target} exists and is not a regular file' ) if os.path.isfile(a_ ): __UpperCamelCase : str =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(F'{download_target} exists, but the SHA256 checksum does not match; re-downloading the file' ) with urllib.request.urlopen(a_ ) as source, open(a_ ,'wb' ) as output: with tqdm( total=int(source.info().get('Content-Length' ) ) ,ncols=80 ,unit='iB' ,unit_scale=a_ ,unit_divisor=1_024 ) as loop: while True: __UpperCamelCase : Optional[Any] =source.read(8_192 ) if not buffer: break output.write(a_ ) loop.update(len(a_ ) ) __UpperCamelCase : List[Any] =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() != expected_shaaaa: raise RuntimeError( 'Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.' ) return model_bytes def A ( a_ ,a_ ) -> Optional[Any]: if ".pt" not in checkpoint_path: __UpperCamelCase : int =_download(_MODELS[checkpoint_path] ) else: __UpperCamelCase : List[str] =torch.load(a_ ,map_location='cpu' ) __UpperCamelCase : Union[str, Any] =original_checkpoint['dims'] __UpperCamelCase : List[Any] =original_checkpoint['model_state_dict'] __UpperCamelCase : Dict =state_dict['decoder.token_embedding.weight'] remove_ignore_keys_(a_ ) rename_keys(a_ ) __UpperCamelCase : List[str] =True __UpperCamelCase : str =state_dict['decoder.layers.0.fc1.weight'].shape[0] __UpperCamelCase : Optional[int] =WhisperConfig( vocab_size=dimensions['n_vocab'] ,encoder_ffn_dim=a_ ,decoder_ffn_dim=a_ ,num_mel_bins=dimensions['n_mels'] ,d_model=dimensions['n_audio_state'] ,max_target_positions=dimensions['n_text_ctx'] ,encoder_layers=dimensions['n_audio_layer'] ,encoder_attention_heads=dimensions['n_audio_head'] ,decoder_layers=dimensions['n_text_layer'] ,decoder_attention_heads=dimensions['n_text_state'] ,max_source_positions=dimensions['n_audio_ctx'] ,) __UpperCamelCase : List[str] =WhisperForConditionalGeneration(a_ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =model.model.load_state_dict(a_ ,strict=a_ ) if len(a_ ) > 0 and not set(a_ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( 'Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,' F' but all the following weights are missing {missing}' ) if tie_embeds: __UpperCamelCase : Optional[int] =make_linear_from_emb(model.model.decoder.embed_tokens ) else: __UpperCamelCase : List[str] =proj_out_weights model.save_pretrained(a_ ) if __name__ == "__main__": A_ :List[Any] = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') A_ :List[Any] = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
71
A_ :str = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
71
1
import sys A_ :List[Any] = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def A ( a_ = N ) -> int: __UpperCamelCase : List[str] =-sys.maxsize - 1 for i in range(len(a_ ) - 12 ): __UpperCamelCase : Optional[int] =1 for j in range(13 ): product *= int(n[i + j] ) if product > largest_product: __UpperCamelCase : List[str] =product return largest_product if __name__ == "__main__": print(f"{solution() = }")
71
A_ :Union[str, Any] = { 0: '''0''', 1: '''1''', 2: '''2''', 3: '''3''', 4: '''4''', 5: '''5''', 6: '''6''', 7: '''7''', 8: '''8''', 9: '''9''', 10: '''a''', 11: '''b''', 12: '''c''', 13: '''d''', 14: '''e''', 15: '''f''', } def A ( a_ ) -> str: assert type(a_ ) in (int, float) and decimal == int(a_ ) __UpperCamelCase : Union[str, Any] =int(a_ ) __UpperCamelCase : List[str] ='' __UpperCamelCase : Optional[Any] =False if decimal < 0: __UpperCamelCase : Tuple =True decimal *= -1 while decimal > 0: __UpperCamelCase , __UpperCamelCase : Optional[Any] =divmod(a_ ,16 ) __UpperCamelCase : Tuple =values[remainder] + hexadecimal __UpperCamelCase : Dict ='0x' + hexadecimal if negative: __UpperCamelCase : int ='-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
71
1
A_ :str = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
71
import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness A_ :List[str] = '''\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } ''' A_ :Any = '''\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). ''' A_ :Tuple = ''' Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric("code_eval") >>> test_cases = ["assert add(2,3)==5"] >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {\'pass@1\': 0.5, \'pass@2\': 1.0} ''' A_ :List[str] = ''' ################################################################################ !!!WARNING!!! ################################################################################ The "code_eval" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this with: >>> import os >>> os.environ["HF_ALLOW_CODE_EVAL"] = "1" ################################################################################\ ''' A_ :Tuple = '''The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): """simple docstring""" def __lowercase ( self ): """simple docstring""" return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' ) ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/openai/human-eval' , codebase_urls=['https://github.com/openai/human-eval'] , reference_urls=['https://github.com/openai/human-eval'] , license=_LICENSE , ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=[1, 10, 100] , lowerCamelCase__=4 , lowerCamelCase__=3.0 ): """simple docstring""" if os.getenv('HF_ALLOW_CODE_EVAL' , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError('This metric is currently not supported on Windows.' ) with ThreadPoolExecutor(max_workers=lowerCamelCase__ ) as executor: __UpperCamelCase : List[str] =[] __UpperCamelCase : Any =Counter() __UpperCamelCase : List[Any] =0 __UpperCamelCase : int =defaultdict(lowerCamelCase__ ) for task_id, (candidates, test_case) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): for candidate in candidates: __UpperCamelCase : str =candidate + '\n' + test_case __UpperCamelCase : Any =(test_program, timeout, task_id, completion_id[task_id]) __UpperCamelCase : Optional[Any] =executor.submit(lowerCamelCase__ , *lowerCamelCase__ ) futures.append(lowerCamelCase__ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(lowerCamelCase__ ): __UpperCamelCase : str =future.result() results[result["task_id"]].append((result['completion_id'], result) ) __UpperCamelCase , __UpperCamelCase : int =[], [] for result in results.values(): result.sort() __UpperCamelCase : str =[r[1]['passed'] for r in result] total.append(len(lowerCamelCase__ ) ) correct.append(sum(lowerCamelCase__ ) ) __UpperCamelCase : Optional[int] =np.array(lowerCamelCase__ ) __UpperCamelCase : List[str] =np.array(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =k __UpperCamelCase : List[Any] ={f'pass@{k}': estimate_pass_at_k(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def A ( a_ ,a_ ,a_ ) -> Optional[int]: def estimator(a_ ,a_ ,a_ ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 ,n + 1 ) ) if isinstance(a_ ,a_ ): __UpperCamelCase : Optional[int] =itertools.repeat(a_ ,len(a_ ) ) else: assert len(a_ ) == len(a_ ) __UpperCamelCase : List[Any] =iter(a_ ) return np.array([estimator(int(a_ ) ,int(a_ ) ,a_ ) for n, c in zip(a_ ,a_ )] )
71
1
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __A ( a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : str =GPTSanJapaneseTokenizer UpperCamelCase__ : Tuple =False UpperCamelCase__ : Dict ={"""do_clean_text""": False, """add_prefix_space""": False} def __lowercase ( self ): """simple docstring""" super().setUp() # fmt: off __UpperCamelCase : Tuple =['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on __UpperCamelCase : str ={'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 __UpperCamelCase : Tuple ={'unk_token': '<unk>'} __UpperCamelCase : Union[str, Any] =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase : Optional[Any] =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(lowerCamelCase__ ) ) def __lowercase ( self , **lowerCamelCase__ ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] ='こんにちは、世界。 \nこんばんは、㔺界。😀' __UpperCamelCase : Dict ='こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Optional[Any] =self.get_input_output_texts(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =tokenizer.decode(lowerCamelCase__ , clean_up_tokenization_spaces=lowerCamelCase__ ) return text, ids def __lowercase ( self ): """simple docstring""" pass # TODO add if relevant def __lowercase ( self ): """simple docstring""" pass # TODO add if relevant def __lowercase ( self ): """simple docstring""" pass # TODO add if relevant def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =self.get_tokenizer() # Testing tokenization __UpperCamelCase : Optional[int] ='こんにちは、世界。 こんばんは、㔺界。' __UpperCamelCase : Tuple =['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] __UpperCamelCase : Any =tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids without special tokens __UpperCamelCase : List[Any] =[0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] __UpperCamelCase : Union[str, Any] =tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids with special tokens __UpperCamelCase : Optional[int] =tokens + [tokenizer.unk_token] __UpperCamelCase : Union[str, Any] =[0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] __UpperCamelCase : Union[str, Any] =tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =self.get_tokenizer() # Testing tokenization __UpperCamelCase : Optional[Any] ='こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' __UpperCamelCase : Dict ='こんにちは、、、、世界。こんばんは、、、、世界。' __UpperCamelCase : str =tokenizer.encode(lowerCamelCase__ ) __UpperCamelCase : Optional[int] =tokenizer.decode(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization __UpperCamelCase : Optional[int] ='こんにちは、世界。' __UpperCamelCase : Optional[Any] ='こんばんは、㔺界。😀' __UpperCamelCase : str ='こんにちは、世界。こんばんは、世界。😀' __UpperCamelCase : Dict =tokenizer.encode(prefix_text + input_text ) __UpperCamelCase : Tuple =tokenizer.encode('' , prefix_text=prefix_text + input_text ) __UpperCamelCase : Union[str, Any] =tokenizer.encode(lowerCamelCase__ , prefix_text=lowerCamelCase__ ) __UpperCamelCase : List[Any] =tokenizer.decode(lowerCamelCase__ ) __UpperCamelCase : Any =tokenizer.decode(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =tokenizer.decode(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization __UpperCamelCase : Union[str, Any] ='こんにちは、世界。' __UpperCamelCase : str ='こんばんは、㔺界。😀' __UpperCamelCase : List[str] =len(tokenizer.encode(lowerCamelCase__ ) ) - 2 __UpperCamelCase : Union[str, Any] =len(tokenizer.encode(lowerCamelCase__ ) ) - 2 __UpperCamelCase : Optional[Any] =[1] + [0] * (len_prefix + len_text + 1) __UpperCamelCase : List[str] =[1] * (len_prefix + len_text + 1) + [0] __UpperCamelCase : Tuple =[1] + [1] * (len_prefix) + [0] * (len_text + 1) __UpperCamelCase : List[Any] =tokenizer(prefix_text + input_text ).token_type_ids __UpperCamelCase : List[Any] =tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids __UpperCamelCase : Any =tokenizer(lowerCamelCase__ , prefix_text=lowerCamelCase__ ).token_type_ids self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) __UpperCamelCase : Optional[int] =tokenizer.encode('あンいワ' ) __UpperCamelCase : Optional[int] =tokenizer.encode('' , prefix_text='あンいワ' ) __UpperCamelCase : Union[str, Any] =tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(lowerCamelCase__ ) , tokenizer.decode(lowerCamelCase__ ) ) self.assertEqual(tokenizer.decode(lowerCamelCase__ ) , tokenizer.decode(lowerCamelCase__ ) ) self.assertNotEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertNotEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) __UpperCamelCase : Union[str, Any] =[['武田信玄', 'は、'], ['織田信長', 'の配下の、']] __UpperCamelCase : str =tokenizer(lowerCamelCase__ , padding=lowerCamelCase__ ) __UpperCamelCase : List[str] =tokenizer.batch_encode_plus(lowerCamelCase__ , padding=lowerCamelCase__ ) # fmt: off __UpperCamelCase : Tuple =[[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] __UpperCamelCase : Union[str, Any] =[[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] __UpperCamelCase : Dict =[[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , lowerCamelCase__ ) self.assertListEqual(x_token.token_type_ids , lowerCamelCase__ ) self.assertListEqual(x_token.attention_mask , lowerCamelCase__ ) self.assertListEqual(x_token_a.input_ids , lowerCamelCase__ ) self.assertListEqual(x_token_a.token_type_ids , lowerCamelCase__ ) self.assertListEqual(x_token_a.attention_mask , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" pass def __lowercase ( self ): """simple docstring""" pass
71
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __A ( a , a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Optional[Any] =StableDiffusionDiffEditPipeline UpperCamelCase__ : str =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""height""", """width""", """image"""} | {"""image_latents"""} UpperCamelCase__ : Optional[Any] =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"""image"""} | {"""image_latents"""} UpperCamelCase__ : Dict =frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess UpperCamelCase__ : Any =frozenset([] ) def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Dict =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , ) __UpperCamelCase : List[str] =DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) __UpperCamelCase : Union[str, Any] =DDIMInverseScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_zero=lowerCamelCase__ , ) torch.manual_seed(0 ) __UpperCamelCase : Optional[int] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __UpperCamelCase : Tuple =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) __UpperCamelCase : Any =CLIPTextModel(lowerCamelCase__ ) __UpperCamelCase : int =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __UpperCamelCase : Union[str, Any] ={ 'unet': unet, 'scheduler': scheduler, 'inverse_scheduler': inverse_scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : int =floats_tensor((1, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Optional[int] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Dict ={ 'prompt': 'a dog and a newt', 'mask_image': mask, 'image_latents': latents, 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : Tuple =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : int =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : Optional[Any] =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : List[Any] =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Any =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'source_prompt': 'a cat and a frog', 'target_prompt': 'a dog and a newt', 'generator': generator, 'num_inference_steps': 2, 'num_maps_per_mask': 2, 'mask_encode_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : str =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : int =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : int =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'prompt': 'a cat and a frog', 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'decode_latents': True, 'output_type': 'numpy', } return inputs def __lowercase ( self ): """simple docstring""" if not hasattr(self.pipeline_class , '_optional_components' ): return __UpperCamelCase : Optional[Any] =self.get_dummy_components() __UpperCamelCase : List[str] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) __UpperCamelCase : Union[str, Any] =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe(**lowerCamelCase__ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : Tuple =self.pipeline_class.from_pretrained(lowerCamelCase__ ) pipe_loaded.to(lowerCamelCase__ ) pipe_loaded.set_progress_bar_config(disable=lowerCamelCase__ ) for optional_component in pipe._optional_components: self.assertTrue( getattr(lowerCamelCase__ , lowerCamelCase__ ) is None , f'`{optional_component}` did not stay set to None after loading.' , ) __UpperCamelCase : str =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe_loaded(**lowerCamelCase__ )[0] __UpperCamelCase : Tuple =np.abs(output - output_loaded ).max() self.assertLess(lowerCamelCase__ , 1E-4 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : int =self.get_dummy_mask_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe.generate_mask(**lowerCamelCase__ ) __UpperCamelCase : int =mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) __UpperCamelCase : Tuple =np.array([0] * 9 ) __UpperCamelCase : str =np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Optional[Any] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Dict =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : Optional[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : int =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def __lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=5E-3 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ='cpu' __UpperCamelCase : int =self.get_dummy_components() __UpperCamelCase : str ={'beta_start': 0.00_085, 'beta_end': 0.012, 'beta_schedule': 'scaled_linear'} __UpperCamelCase : str =DPMSolverMultistepScheduler(**lowerCamelCase__ ) __UpperCamelCase : Dict =DPMSolverMultistepInverseScheduler(**lowerCamelCase__ ) __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : str =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : List[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : Optional[Any] =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) @require_torch_gpu @slow class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def __lowercase ( cls ): """simple docstring""" __UpperCamelCase : Optional[int] =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png' ) __UpperCamelCase : Union[str, Any] =raw_image.convert('RGB' ).resize((768, 768) ) __UpperCamelCase : List[Any] =raw_image def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =torch.manual_seed(0 ) __UpperCamelCase : Dict =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : List[str] =DDIMScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : List[str] =DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : List[str] ='a bowl of fruit' __UpperCamelCase : Dict ='a bowl of pears' __UpperCamelCase : Tuple =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : int =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ ).latents __UpperCamelCase : Dict =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , output_type='numpy' , ).images[0] __UpperCamelCase : str =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =torch.manual_seed(0 ) __UpperCamelCase : List[Any] =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : Optional[Any] =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : Optional[int] =DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] ='a bowl of fruit' __UpperCamelCase : int ='a bowl of pears' __UpperCamelCase : str =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : List[str] =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ , num_inference_steps=25 , ).latents __UpperCamelCase : List[str] =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , num_inference_steps=25 , output_type='numpy' , ).images[0] __UpperCamelCase : Tuple =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1
71
1
A_ :dict[tuple[int, int, int], int] = {} def A ( a_ ,a_ ,a_ ) -> int: # if we are absent twice, or late 3 consecutive days, # no further prize strings are possible if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on __UpperCamelCase : List[str] =(days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one __UpperCamelCase : Any =_calculate(days - 1 ,a_ ,late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 __UpperCamelCase : List[Any] =_calculate(days - 1 ,absent + 1 ,0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter __UpperCamelCase : List[Any] =_calculate(days - 1 ,a_ ,0 ) __UpperCamelCase : Optional[int] =state_late + state_absent + state_ontime __UpperCamelCase : Any =prizestrings return prizestrings def A ( a_ = 30 ) -> int: return _calculate(a_ ,absent=0 ,late=0 ) if __name__ == "__main__": print(solution())
71
import random from .binary_exp_mod import bin_exp_mod def A ( a_ ,a_=1_000 ) -> Optional[Any]: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd __UpperCamelCase : List[Any] =n - 1 __UpperCamelCase : Dict =0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) __UpperCamelCase : Optional[Any] =0 while count < prec: __UpperCamelCase : Dict =random.randint(2 ,n - 1 ) __UpperCamelCase : Optional[Any] =bin_exp_mod(a_ ,a_ ,a_ ) if b != 1: __UpperCamelCase : List[str] =True for _ in range(a_ ): if b == n - 1: __UpperCamelCase : Tuple =False break __UpperCamelCase : Dict =b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": A_ :str = abs(int(input('''Enter bound : ''').strip())) print('''Here\'s the list of primes:''') print(''', '''.join(str(i) for i in range(n + 1) if is_prime_big(i)))
71
1
from __future__ import annotations import requests def A ( a_ ) -> dict: __UpperCamelCase : int =F'https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty' return requests.get(a_ ).json() def A ( a_ = 10 ) -> list[dict]: __UpperCamelCase : str ='https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty' __UpperCamelCase : List[str] =requests.get(a_ ).json()[:max_stories] return [get_hackernews_story(a_ ) for story_id in story_ids] def A ( a_ = 10 ) -> str: __UpperCamelCase : List[str] =hackernews_top_stories(a_ ) return "\n".join('* [{title}]({url})'.format(**a_ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
71
from torch import nn class __A ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" super().__init__() __UpperCamelCase : Dict =class_size __UpperCamelCase : Any =embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) __UpperCamelCase : Any =nn.Linear(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : List[Any] =self.mlp(lowerCamelCase__ ) return logits
71
1
import fire from utils import calculate_rouge, save_json def A ( a_ ,a_ ,a_=None ,**a_ ) -> Tuple: __UpperCamelCase : Dict =[x.strip() for x in open(a_ ).readlines()] __UpperCamelCase : str =[x.strip() for x in open(a_ ).readlines()][: len(a_ )] __UpperCamelCase : Union[str, Any] =calculate_rouge(a_ ,a_ ,**a_ ) if save_path is not None: save_json(a_ ,a_ ,indent=a_ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
71
def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square(a_ ,a_ ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __UpperCamelCase : Optional[int] =update_area_of_max_square(a_ ,col + 1 ) __UpperCamelCase : List[str] =update_area_of_max_square(row + 1 ,col + 1 ) __UpperCamelCase : List[Any] =update_area_of_max_square(row + 1 ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : Dict =max(largest_square_area[0] ,a_ ) return sub_problem_sol else: return 0 __UpperCamelCase : Union[str, Any] =[0] update_area_of_max_square(0 ,0 ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square_using_dp_array( a_ ,a_ ,a_ ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __UpperCamelCase : Tuple =update_area_of_max_square_using_dp_array(a_ ,col + 1 ,a_ ) __UpperCamelCase : Optional[int] =update_area_of_max_square_using_dp_array(row + 1 ,col + 1 ,a_ ) __UpperCamelCase : Any =update_area_of_max_square_using_dp_array(row + 1 ,a_ ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : str =max(largest_square_area[0] ,a_ ) __UpperCamelCase : Any =sub_problem_sol return sub_problem_sol else: return 0 __UpperCamelCase : Tuple =[0] __UpperCamelCase : List[Any] =[[-1] * cols for _ in range(a_ )] update_area_of_max_square_using_dp_array(0 ,0 ,a_ ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Dict =[[0] * (cols + 1) for _ in range(rows + 1 )] __UpperCamelCase : int =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Optional[Any] =dp_array[row][col + 1] __UpperCamelCase : int =dp_array[row + 1][col + 1] __UpperCamelCase : Tuple =dp_array[row + 1][col] if mat[row][col] == 1: __UpperCamelCase : Tuple =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Any =max(dp_array[row][col] ,a_ ) else: __UpperCamelCase : Dict =0 return largest_square_area def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Any =[0] * (cols + 1) __UpperCamelCase : List[Any] =[0] * (cols + 1) __UpperCamelCase : Tuple =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Any =current_row[col + 1] __UpperCamelCase : Optional[Any] =next_row[col + 1] __UpperCamelCase : Union[str, Any] =next_row[col] if mat[row][col] == 1: __UpperCamelCase : Any =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Optional[int] =max(current_row[col] ,a_ ) else: __UpperCamelCase : List[str] =0 __UpperCamelCase : Optional[Any] =current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
71
1
from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize('repo_id' ,['canonical_dataset_name', 'org-name/dataset-name'] ) @pytest.mark.parametrize('path' ,['filename.csv', 'filename with blanks.csv'] ) @pytest.mark.parametrize('revision' ,[None, 'v2'] ) def A ( a_ ,a_ ,a_ ) -> Optional[int]: __UpperCamelCase : Tuple =hf_hub_url(repo_id=a_ ,path=a_ ,revision=a_ ) assert url == F'https://huggingface.co/datasets/{repo_id}/resolve/{revision or "main"}/{quote(a_ )}'
71
def A ( a_ ) -> int: __UpperCamelCase : Any =len(a_ ) while cur > 1: # Find the maximum number in arr __UpperCamelCase : Any =arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi __UpperCamelCase : Any =arr[mi::-1] + arr[mi + 1 : len(a_ )] # Reverse whole list __UpperCamelCase : str =arr[cur - 1 :: -1] + arr[cur : len(a_ )] cur -= 1 return arr if __name__ == "__main__": A_ :Dict = input('''Enter numbers separated by a comma:\n''').strip() A_ :Any = [int(item) for item in user_input.split(''',''')] print(pancake_sort(unsorted))
71
1
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=a ) class __A ( a ): """simple docstring""" UpperCamelCase__ : str =field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) UpperCamelCase__ : ClassVar[Features] =Features({"""audio""": Audio()} ) UpperCamelCase__ : ClassVar[Features] =Features({"""transcription""": Value("""string""" )} ) UpperCamelCase__ : str ="audio" UpperCamelCase__ : str ="transcription" def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" if self.audio_column not in features: raise ValueError(f'Column {self.audio_column} is not present in features.' ) if not isinstance(features[self.audio_column] , lowerCamelCase__ ): raise ValueError(f'Column {self.audio_column} is not an Audio type.' ) __UpperCamelCase : Optional[int] =copy.deepcopy(self ) __UpperCamelCase : int =self.input_schema.copy() __UpperCamelCase : Union[str, Any] =features[self.audio_column] __UpperCamelCase : List[Any] =input_schema return task_template @property def __lowercase ( self ): """simple docstring""" return {self.audio_column: "audio", self.transcription_column: "transcription"}
71
import random def A ( a_ ,a_ ,a_ = False ) -> dict: __UpperCamelCase : dict ={i: [] for i in range(a_ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(a_ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(a_ ): for j in range(i + 1 ,a_ ): if random.random() < probability: graph[i].append(a_ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(a_ ) return graph def A ( a_ ) -> dict: return { i: [j for j in range(a_ ) if i != j] for i in range(a_ ) } if __name__ == "__main__": import doctest doctest.testmod()
71
1
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin A_ :Dict = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class __A ( a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : str =XLMProphetNetTokenizer UpperCamelCase__ : int =False UpperCamelCase__ : Optional[int] =True def __lowercase ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCamelCase : List[Any] =XLMProphetNetTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] ='[PAD]' __UpperCamelCase : Union[str, Any] =0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '[PAD]' ) self.assertEqual(vocab_keys[1] , '[CLS]' ) self.assertEqual(vocab_keys[-1] , 'j' ) self.assertEqual(len(lowerCamelCase__ ) , 1012 ) def __lowercase ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =XLMProphetNetTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) __UpperCamelCase : List[str] =tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __UpperCamelCase : List[str] =tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __UpperCamelCase : List[str] =tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) __UpperCamelCase : int =tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '[UNK]', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '[UNK]', '.', ] , ) @cached_property def __lowercase ( self ): """simple docstring""" return XLMProphetNetTokenizer.from_pretrained('microsoft/xprophetnet-large-wiki100-cased' ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str ='Hello World!' __UpperCamelCase : Union[str, Any] =[35389, 6672, 49, 2] self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int ={'input_ids': [[11073, 82783, 18, 26, 82783, 549, 51540, 248, 17209, 1301, 217, 20, 215186, 1325, 147, 17209, 1301, 217, 20, 56370, 53, 122020, 20, 16477, 27, 87355, 4548, 20, 4728, 78392, 17, 159969, 18, 26, 24491, 629, 15, 538, 22704, 5439, 15, 2788, 24491, 9885, 15, 43534, 605, 15, 814, 18403, 33200, 29, 15, 43534, 24458, 12410, 111, 24966, 83669, 9637, 144068, 26, 850, 22346, 27, 147, 24966, 83669, 83490, 26, 39113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 122020, 115785, 34, 816, 1339, 46887, 18, 147, 53905, 1951, 42238, 41170, 17732, 834, 436, 15, 27523, 98733, 217, 147, 5542, 4981, 930, 17347, 16, 2], [20091, 629, 94, 82786, 58, 490, 20, 1528, 84, 53905, 344, 80592, 110128, 18822, 5267, 1306, 62, 152537, 308, 7997, 401, 124427, 549, 35442, 225, 109, 15055, 25748, 147, 7119, 43712, 34, 767, 135366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63784, 119466, 17, 147808, 88214, 18, 656, 81, 32, 3296, 10280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='microsoft/xprophetnet-large-wiki100-cased' , revision='1acad1643ddd54a44df6a1b797ada8373685d90e' , )
71
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[str] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModel.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =AutoModel.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForPreTraining.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =AutoModelForPreTraining.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Tuple =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Tuple =TFAutoModelForCausalLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =AutoModelForCausalLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =AutoModelForCausalLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : int =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Dict =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Tuple =TFAutoModelForMaskedLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =AutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Optional[Any] =AutoModelForMaskedLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[str] =TFAutoModelForSeqaSeqLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Dict =AutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[Any] =AutoModelForSeqaSeqLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : str =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =TFAutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =AutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =TFAutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =AutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) __UpperCamelCase : str =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) __UpperCamelCase : int =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 )
71
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def A ( a_ ) -> int: __UpperCamelCase : List[str] =384 if "tiny" in model_name: __UpperCamelCase : Tuple =[3, 3, 9, 3] __UpperCamelCase : Union[str, Any] =[96, 192, 384, 768] if "small" in model_name: __UpperCamelCase : Optional[Any] =[3, 3, 27, 3] __UpperCamelCase : Dict =[96, 192, 384, 768] if "base" in model_name: __UpperCamelCase : Union[str, Any] =[3, 3, 27, 3] __UpperCamelCase : Any =[128, 256, 512, 1_024] __UpperCamelCase : List[str] =512 if "large" in model_name: __UpperCamelCase : Union[str, Any] =[3, 3, 27, 3] __UpperCamelCase : Tuple =[192, 384, 768, 1_536] __UpperCamelCase : str =768 if "xlarge" in model_name: __UpperCamelCase : List[Any] =[3, 3, 27, 3] __UpperCamelCase : Dict =[256, 512, 1_024, 2_048] __UpperCamelCase : int =1_024 # set label information __UpperCamelCase : int =150 __UpperCamelCase : List[str] ='huggingface/label-files' __UpperCamelCase : int ='ade20k-id2label.json' __UpperCamelCase : Union[str, Any] =json.load(open(hf_hub_download(a_ ,a_ ,repo_type='dataset' ) ,'r' ) ) __UpperCamelCase : Any ={int(a_ ): v for k, v in idalabel.items()} __UpperCamelCase : List[str] ={v: k for k, v in idalabel.items()} __UpperCamelCase : Optional[Any] =ConvNextConfig( depths=a_ ,hidden_sizes=a_ ,out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) __UpperCamelCase : int =UperNetConfig( backbone_config=a_ ,auxiliary_in_channels=a_ ,num_labels=a_ ,idalabel=a_ ,labelaid=a_ ,) return config def A ( a_ ) -> str: __UpperCamelCase : Tuple =[] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'backbone.stages.{i}.{j}.gamma', F'backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.weight', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.depthwise_conv.bias', F'backbone.encoder.stages.{i}.layers.{j}.dwconv.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.weight', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.norm.bias', F'backbone.encoder.stages.{i}.layers.{j}.layernorm.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv1.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.weight', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight') ) rename_keys.append((F'backbone.stages.{i}.{j}.pointwise_conv2.bias', F'backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias') ) if i > 0: rename_keys.append((F'backbone.downsample_layers.{i}.0.weight', F'backbone.encoder.stages.{i}.downsampling_layer.0.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.0.bias', F'backbone.encoder.stages.{i}.downsampling_layer.0.bias') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.weight', F'backbone.encoder.stages.{i}.downsampling_layer.1.weight') ) rename_keys.append((F'backbone.downsample_layers.{i}.1.bias', F'backbone.encoder.stages.{i}.downsampling_layer.1.bias') ) rename_keys.append((F'backbone.norm{i}.weight', F'backbone.hidden_states_norms.stage{i+1}.weight') ) rename_keys.append((F'backbone.norm{i}.bias', F'backbone.hidden_states_norms.stage{i+1}.bias') ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def A ( a_ ,a_ ,a_ ) -> Optional[Any]: __UpperCamelCase : int =dct.pop(a_ ) __UpperCamelCase : Optional[Any] =val def A ( a_ ,a_ ,a_ ) -> Optional[int]: __UpperCamelCase : Tuple ={ 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } __UpperCamelCase : Optional[Any] =model_name_to_url[model_name] __UpperCamelCase : str =torch.hub.load_state_dict_from_url(a_ ,map_location='cpu' )['state_dict'] __UpperCamelCase : Dict =get_upernet_config(a_ ) __UpperCamelCase : Tuple =UperNetForSemanticSegmentation(a_ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): __UpperCamelCase : Any =state_dict.pop(a_ ) if "bn" in key: __UpperCamelCase : List[Any] =key.replace('bn' ,'batch_norm' ) __UpperCamelCase : int =val # rename keys __UpperCamelCase : List[Any] =create_rename_keys(a_ ) for src, dest in rename_keys: rename_key(a_ ,a_ ,a_ ) model.load_state_dict(a_ ) # verify on image __UpperCamelCase : List[Any] ='https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' __UpperCamelCase : Union[str, Any] =Image.open(requests.get(a_ ,stream=a_ ).raw ).convert('RGB' ) __UpperCamelCase : Union[str, Any] =SegformerImageProcessor() __UpperCamelCase : Optional[Any] =processor(a_ ,return_tensors='pt' ).pixel_values with torch.no_grad(): __UpperCamelCase : List[Any] =model(a_ ) if model_name == "upernet-convnext-tiny": __UpperCamelCase : List[str] =torch.tensor( [[-8.8_110, -8.8_110, -8.6_521], [-8.8_110, -8.8_110, -8.6_521], [-8.7_746, -8.7_746, -8.6_130]] ) elif model_name == "upernet-convnext-small": __UpperCamelCase : Optional[int] =torch.tensor( [[-8.8_236, -8.8_236, -8.6_771], [-8.8_236, -8.8_236, -8.6_771], [-8.7_638, -8.7_638, -8.6_240]] ) elif model_name == "upernet-convnext-base": __UpperCamelCase : Union[str, Any] =torch.tensor( [[-8.8_558, -8.8_558, -8.6_905], [-8.8_558, -8.8_558, -8.6_905], [-8.7_669, -8.7_669, -8.6_021]] ) elif model_name == "upernet-convnext-large": __UpperCamelCase : Optional[int] =torch.tensor( [[-8.6_660, -8.6_660, -8.6_210], [-8.6_660, -8.6_660, -8.6_210], [-8.6_310, -8.6_310, -8.5_964]] ) elif model_name == "upernet-convnext-xlarge": __UpperCamelCase : Any =torch.tensor( [[-8.4_980, -8.4_980, -8.3_977], [-8.4_980, -8.4_980, -8.3_977], [-8.4_379, -8.4_379, -8.3_412]] ) print('Logits:' ,outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] ,a_ ,atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(a_ ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(a_ ) if push_to_hub: print(F'Pushing model and processor for {model_name} to hub' ) model.push_to_hub(F'openmmlab/{model_name}' ) processor.push_to_hub(F'openmmlab/{model_name}' ) if __name__ == "__main__": A_ :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''upernet-convnext-tiny''', type=str, choices=[f"upernet-convnext-{size}" for size in ['''tiny''', '''small''', '''base''', '''large''', '''xlarge''']], help='''Name of the ConvNext UperNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ :Any = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
71
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ :Tuple = { '''configuration_x_clip''': [ '''XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XCLIPConfig''', '''XCLIPTextConfig''', '''XCLIPVisionConfig''', ], '''processing_x_clip''': ['''XCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Union[str, Any] = [ '''XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XCLIPModel''', '''XCLIPPreTrainedModel''', '''XCLIPTextModel''', '''XCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys A_ :Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
71
1
def A ( a_ = 1_000 ) -> int: __UpperCamelCase : List[Any] =2**power __UpperCamelCase : Optional[int] =str(a_ ) __UpperCamelCase : Any =list(a_ ) __UpperCamelCase : Tuple =0 for i in list_num: sum_of_num += int(a_ ) return sum_of_num if __name__ == "__main__": A_ :Optional[int] = int(input('''Enter the power of 2: ''').strip()) print('''2 ^ ''', power, ''' = ''', 2**power) A_ :Tuple = solution(power) print('''Sum of the digits is: ''', result)
71
from ...configuration_utils import PretrainedConfig from ...utils import logging A_ :Any = logging.get_logger(__name__) A_ :int = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[int] ="""vit_msn""" def __init__( self , lowerCamelCase__=768 , lowerCamelCase__=12 , lowerCamelCase__=12 , lowerCamelCase__=3072 , lowerCamelCase__="gelu" , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.02 , lowerCamelCase__=1E-06 , lowerCamelCase__=224 , lowerCamelCase__=16 , lowerCamelCase__=3 , lowerCamelCase__=True , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : int =hidden_size __UpperCamelCase : List[Any] =num_hidden_layers __UpperCamelCase : Union[str, Any] =num_attention_heads __UpperCamelCase : List[str] =intermediate_size __UpperCamelCase : Union[str, Any] =hidden_act __UpperCamelCase : str =hidden_dropout_prob __UpperCamelCase : Union[str, Any] =attention_probs_dropout_prob __UpperCamelCase : Union[str, Any] =initializer_range __UpperCamelCase : Tuple =layer_norm_eps __UpperCamelCase : Optional[Any] =image_size __UpperCamelCase : Optional[int] =patch_size __UpperCamelCase : Any =num_channels __UpperCamelCase : str =qkv_bias
71
1
import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A ( a_ ,a_ ,a_ ) -> Tuple: # Construct model if openai_config_file == "": __UpperCamelCase : int =OpenAIGPTConfig() else: __UpperCamelCase : Any =OpenAIGPTConfig.from_json_file(a_ ) __UpperCamelCase : Any =OpenAIGPTModel(a_ ) # Load weights from numpy load_tf_weights_in_openai_gpt(a_ ,a_ ,a_ ) # Save pytorch-model __UpperCamelCase : Optional[Any] =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase : Optional[Any] =pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() ,a_ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A_ :int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--openai_checkpoint_folder_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--openai_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained OpenAI model. \n''' '''This specifies the model architecture.''' ), ) A_ :Optional[int] = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
71
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __A ( a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : str =DDIMPipeline UpperCamelCase__ : List[Any] =UNCONDITIONAL_IMAGE_GENERATION_PARAMS UpperCamelCase__ : Tuple =PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } UpperCamelCase__ : Tuple =UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS UpperCamelCase__ : Any =False def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Optional[int] =UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) __UpperCamelCase : int =DDIMScheduler() __UpperCamelCase : Optional[int] ={'unet': unet, 'scheduler': scheduler} return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : str =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Optional[int] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Tuple ={ 'batch_size': 1, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='cpu' __UpperCamelCase : Optional[Any] =self.get_dummy_components() __UpperCamelCase : Tuple =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : int =pipe(**lowerCamelCase__ ).images __UpperCamelCase : Dict =image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) __UpperCamelCase : Tuple =np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) __UpperCamelCase : Tuple =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def __lowercase ( self ): """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str ='google/ddpm-cifar10-32' __UpperCamelCase : str =UNetaDModel.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =DDIMScheduler() __UpperCamelCase : List[Any] =DDIMPipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) ddim.to(lowerCamelCase__ ) ddim.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[int] =torch.manual_seed(0 ) __UpperCamelCase : List[str] =ddim(generator=lowerCamelCase__ , eta=0.0 , output_type='numpy' ).images __UpperCamelCase : Union[str, Any] =image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase : str =np.array([0.1_723, 0.1_617, 0.1_600, 0.1_626, 0.1_497, 0.1_513, 0.1_505, 0.1_442, 0.1_453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] ='google/ddpm-ema-bedroom-256' __UpperCamelCase : Any =UNetaDModel.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : int =DDIMScheduler.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : Dict =DDIMPipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) ddpm.to(lowerCamelCase__ ) ddpm.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple =torch.manual_seed(0 ) __UpperCamelCase : Union[str, Any] =ddpm(generator=lowerCamelCase__ , output_type='numpy' ).images __UpperCamelCase : Tuple =image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __UpperCamelCase : Optional[Any] =np.array([0.0_060, 0.0_201, 0.0_344, 0.0_024, 0.0_018, 0.0_002, 0.0_022, 0.0_000, 0.0_069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
71
1
from typing import Union import fire import torch from tqdm import tqdm def A ( a_ ,a_ = "cpu" ,a_ = None ) -> None: __UpperCamelCase : Tuple =torch.load(a_ ,map_location=a_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(a_ ,torch.Tensor ): raise TypeError('FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin' ) __UpperCamelCase : Optional[int] =v.half() if save_path is None: # overwrite src_path __UpperCamelCase : int =src_path torch.save(a_ ,a_ ) if __name__ == "__main__": fire.Fire(convert)
71
from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[int] ="""new-model""" if is_tf_available(): class __A ( a ): """simple docstring""" UpperCamelCase__ : List[str] =NewModelConfig @require_tf class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ='bert-base-cased' __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] ='bert-base-cased' __UpperCamelCase : Optional[int] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =TFAutoModelForPreTraining.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : str =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : str =TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Optional[int] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Any =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Any =TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : str =TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : Union[str, Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =TFAutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow @require_tensorflow_probability def __lowercase ( self ): """simple docstring""" for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModelForTableQuestionAnswering.from_pretrained(lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[str] =TFAutoModelForTableQuestionAnswering.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =copy.deepcopy(model.config ) __UpperCamelCase : Optional[Any] =['FunnelBaseModel'] __UpperCamelCase : Tuple =TFAutoModel.from_config(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : List[Any] =TFAutoModel.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" try: AutoConfig.register('new-model' , lowerCamelCase__ ) __UpperCamelCase : int =[ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowerCamelCase__ ): auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCamelCase__ ): auto_class.register(lowerCamelCase__ , lowerCamelCase__ ) # Now that the config is registered, it can be used as any other config with the auto-API __UpperCamelCase : List[str] =BertModelTester(self ).get_config() __UpperCamelCase : Optional[Any] =NewModelConfig(**tiny_config.to_dict() ) __UpperCamelCase : Dict =auto_class.from_config(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =auto_class.from_pretrained(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex( lowerCamelCase__ , 'bert-base is not a local folder and is not a valid model identifier' ): __UpperCamelCase : Dict =TFAutoModel.from_pretrained('bert-base' ) def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex( lowerCamelCase__ , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): __UpperCamelCase : Union[str, Any] =TFAutoModel.from_pretrained(lowerCamelCase__ , revision='aaaaaa' ) def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex( lowerCamelCase__ , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): __UpperCamelCase : List[str] =TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def __lowercase ( self ): """simple docstring""" with self.assertRaisesRegex(lowerCamelCase__ , 'Use `from_pt=True` to load this model' ): __UpperCamelCase : List[Any] =TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: __UpperCamelCase : Dict =TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint __UpperCamelCase : Dict =TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: __UpperCamelCase : Union[str, Any] =TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
71
1
def A ( a_ ) -> int: return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def A ( a_ ) -> bool: __UpperCamelCase : Union[str, Any] =0 __UpperCamelCase : Optional[int] =number while duplicate > 0: __UpperCamelCase , __UpperCamelCase : Any =divmod(a_ ,10 ) fact_sum += factorial(a_ ) return fact_sum == number if __name__ == "__main__": print('''Program to check whether a number is a Krisnamurthy Number or not.''') A_ :Tuple = int(input('''Enter number: ''').strip()) print( f"{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number." )
71
import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() A_ :List[str] = [ '''word_embeddings_layernorm.weight''', '''word_embeddings_layernorm.bias''', '''input_layernorm.weight''', '''input_layernorm.bias''', '''post_attention_layernorm.weight''', '''post_attention_layernorm.bias''', '''self_attention.dense.bias''', '''mlp.dense_4h_to_h.bias''', '''ln_f.weight''', '''ln_f.bias''', ] A_ :Optional[Any] = [ '''mlp.dense_4h_to_h.weight''', '''self_attention.dense.weight''', ] def A ( a_ ,a_ ) -> str: __UpperCamelCase : Any ={ 'word_embeddings.weight': 'word_embeddings.weight', 'word_embeddings.norm.weight': 'word_embeddings_layernorm.weight', 'word_embeddings.norm.bias': 'word_embeddings_layernorm.bias', 'weight': 'ln_f.weight', 'bias': 'ln_f.bias', } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks __UpperCamelCase : Tuple =int(re.match(r'.*layer_(\d*).*' ,a_ )[1] ) layer_number -= 3 return F'h.{layer_number}.' + key def A ( a_ ) -> Any: if dtype == torch.bool: return 1 / 8 __UpperCamelCase : Dict =re.search(r'[^\d](\d+)$' ,str(a_ ) ) if bit_search is None: raise ValueError(F'`dtype` is not a valid dtype: {dtype}.' ) __UpperCamelCase : Tuple =int(bit_search.groups()[0] ) return bit_size // 8 def A ( a_ ,a_ ,a_ ,a_ ,a_ ) -> Dict: # Construct model if bloom_config_file == "": __UpperCamelCase : List[Any] =BloomConfig() else: __UpperCamelCase : List[str] =BloomConfig.from_json_file(a_ ) if shard_model: __UpperCamelCase : int =os.listdir(a_ ) __UpperCamelCase : Union[str, Any] =sorted(filter(lambda a_ : s.startswith('layer' ) and "model_00" in s ,a_ ) ) __UpperCamelCase : Optional[Any] ={'weight_map': {}, 'metadata': {}} __UpperCamelCase : Dict =0 __UpperCamelCase : int =None __UpperCamelCase : Any =BloomConfig() for j, file in enumerate(a_ ): print('Processing file: {}'.format(a_ ) ) __UpperCamelCase : Optional[int] =None for i in range(a_ ): # load all TP files __UpperCamelCase : Dict =file.replace('model_00' ,F'model_0{i}' ) __UpperCamelCase : Optional[Any] =torch.load(os.path.join(a_ ,a_ ) ,map_location='cpu' ) # Rename keys in the transformers names __UpperCamelCase : int =list(temp.keys() ) for key in keys: __UpperCamelCase : Dict =temp.pop(a_ ) if tensors is None: __UpperCamelCase : Any =temp else: for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel __UpperCamelCase : List[Any] =1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks __UpperCamelCase : Any =torch.cat([tensors[key], temp[key]] ,dim=a_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): __UpperCamelCase : Optional[Any] =tensors[key] / pretraining_tp torch.save( a_ ,os.path.join( a_ ,'pytorch_model_{}-of-{}.bin'.format(str(j + 1 ).zfill(5 ) ,str(len(a_ ) ).zfill(5 ) ) ,) ,) for key in tensors.keys(): __UpperCamelCase : Union[str, Any] =tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: __UpperCamelCase : int ='pytorch_model_{}-of-{}.bin'.format( str(j + 1 ).zfill(5 ) ,str(len(a_ ) ).zfill(5 ) ) __UpperCamelCase : Union[str, Any] =BloomConfig() __UpperCamelCase : Tuple =pytorch_dump_folder_path + '/' + CONFIG_NAME __UpperCamelCase : Optional[int] =total_size with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) with open(os.path.join(a_ ,WEIGHTS_NAME + '.index.json' ) ,'w' ,encoding='utf-8' ) as f: __UpperCamelCase : List[Any] =json.dumps(a_ ,indent=2 ,sort_keys=a_ ) + '\n' f.write(a_ ) else: __UpperCamelCase : List[Any] =BloomModel(a_ ) __UpperCamelCase : Optional[Any] =os.listdir(a_ ) __UpperCamelCase : Dict =sorted(filter(lambda a_ : s.startswith('layer' ) and "model_00" in s ,a_ ) ) __UpperCamelCase : Any =None for i, file in enumerate(a_ ): __UpperCamelCase : Union[str, Any] =None for i in range(a_ ): # load all TP files __UpperCamelCase : Optional[Any] =file.replace('model_00' ,F'model_0{i}' ) __UpperCamelCase : str =torch.load(os.path.join(a_ ,a_ ) ,map_location='cpu' ) # Rename keys in the transformers names __UpperCamelCase : List[str] =list(temp.keys() ) for key in keys: __UpperCamelCase : Union[str, Any] =temp.pop(a_ ) if tensors is None: __UpperCamelCase : Optional[Any] =temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel __UpperCamelCase : Optional[int] =1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks __UpperCamelCase : int =torch.cat([tensors[key], temp[key]] ,dim=a_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(a_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): __UpperCamelCase : Dict =tensors[key] / pretraining_tp __UpperCamelCase : str =model.load_state_dict(a_ ,strict=a_ ) assert not other_keys.unexpected_keys, F'The keys {other_keys.unexpected_keys} are unexpected' if missing_keys is None: __UpperCamelCase : str =set(other_keys.missing_keys ) else: __UpperCamelCase : int =missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F'The keys {missing_keys} are missing' # Save pytorch-model os.makedirs(a_ ,exist_ok=a_ ) __UpperCamelCase : Optional[int] =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase : Dict =pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}' ) if config.torch_dtype is not None: __UpperCamelCase : List[str] =model.to(config.torch_dtype ) torch.save(model.state_dict() ,a_ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A_ :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bloom_checkpoint_path''', default=None, type=str, required=True, help='''Path to the Megatron-LM checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--bloom_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--shard_model''', action='''store_true''', help='''An optional setting to shard the output model \nThis enables sharding the converted checkpoint''', ) parser.add_argument( '''--pretraining_tp''', default=4, type=int, help='''Pretraining TP rank that has been used when training the model in Megatron-LM \n''', ) A_ :str = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
71
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __A ( a ): """simple docstring""" UpperCamelCase__ : str =["""image_processor""", """tokenizer"""] UpperCamelCase__ : str ="""ViTImageProcessor""" UpperCamelCase__ : str =("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Tuple =None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , lowerCamelCase__ , ) __UpperCamelCase : int =kwargs.pop('feature_extractor' ) __UpperCamelCase : Dict =image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(lowerCamelCase__ , lowerCamelCase__ ) def __call__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ ): """simple docstring""" if text is None and visual_prompt is None and images is None: raise ValueError('You have to specify either text, visual prompt or images.' ) if text is not None and visual_prompt is not None: raise ValueError('You have to specify exactly one type of prompt. Either text or visual prompt.' ) if text is not None: __UpperCamelCase : Optional[Any] =self.tokenizer(lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ ) if visual_prompt is not None: __UpperCamelCase : str =self.image_processor(lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ ) if images is not None: __UpperCamelCase : List[str] =self.image_processor(lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ ) if visual_prompt is not None and images is not None: __UpperCamelCase : str ={ 'pixel_values': image_features.pixel_values, 'conditional_pixel_values': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: __UpperCamelCase : str =image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: __UpperCamelCase : Tuple ={ 'conditional_pixel_values': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**lowerCamelCase__ ) , tensor_type=lowerCamelCase__ ) def __lowercase ( self , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" return self.tokenizer.batch_decode(*lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" return self.tokenizer.decode(*lowerCamelCase__ , **lowerCamelCase__ ) @property def __lowercase ( self ): """simple docstring""" warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , lowerCamelCase__ , ) return self.image_processor_class @property def __lowercase ( self ): """simple docstring""" warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , lowerCamelCase__ , ) return self.image_processor
71
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class __A : """simple docstring""" UpperCamelCase__ : int =XGLMConfig UpperCamelCase__ : Optional[Any] ={} UpperCamelCase__ : List[str] ="""gelu""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=14 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=99 , lowerCamelCase__=32 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=37 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=512 , lowerCamelCase__=0.02 , ): """simple docstring""" __UpperCamelCase : Tuple =parent __UpperCamelCase : List[str] =batch_size __UpperCamelCase : str =seq_length __UpperCamelCase : Dict =is_training __UpperCamelCase : Tuple =use_input_mask __UpperCamelCase : List[Any] =use_labels __UpperCamelCase : Any =vocab_size __UpperCamelCase : List[Any] =d_model __UpperCamelCase : Optional[int] =num_hidden_layers __UpperCamelCase : List[str] =num_attention_heads __UpperCamelCase : Optional[int] =ffn_dim __UpperCamelCase : str =activation_function __UpperCamelCase : Any =activation_dropout __UpperCamelCase : Optional[int] =attention_dropout __UpperCamelCase : Optional[int] =max_position_embeddings __UpperCamelCase : Any =initializer_range __UpperCamelCase : Dict =None __UpperCamelCase : Optional[int] =0 __UpperCamelCase : Optional[Any] =2 __UpperCamelCase : str =1 def __lowercase ( self ): """simple docstring""" return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) __UpperCamelCase : Union[str, Any] =None if self.use_input_mask: __UpperCamelCase : Dict =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase : Any =self.get_config() __UpperCamelCase : Optional[Any] =floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def __lowercase ( self ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=lowerCamelCase__ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=lowerCamelCase__ , ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) : int =config_and_inputs __UpperCamelCase : Optional[Any] ={ 'input_ids': input_ids, 'head_mask': head_mask, } return config, inputs_dict @require_tf class __A ( a , a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Union[str, Any] =(TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () UpperCamelCase__ : str =(TFXGLMForCausalLM,) if is_tf_available() else () UpperCamelCase__ : Optional[Any] =( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) UpperCamelCase__ : Tuple =False UpperCamelCase__ : Tuple =False UpperCamelCase__ : Optional[Any] =False def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =TFXGLMModelTester(self ) __UpperCamelCase : Dict =ConfigTester(self , config_class=lowerCamelCase__ , n_embd=37 ) def __lowercase ( self ): """simple docstring""" self.config_tester.run_common_tests() @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Optional[Any] =TFXGLMModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def __lowercase ( self ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self , lowerCamelCase__=True ): """simple docstring""" __UpperCamelCase : int =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : List[str] =tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off __UpperCamelCase : str =[2, 268, 9865, 67, 11, 1988, 57252, 9865, 5, 984, 67, 1988, 213838, 1658, 53, 70446, 33, 6657, 278, 1581] # fmt: on __UpperCamelCase : Optional[Any] =model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Union[str, Any] =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) __UpperCamelCase : str =tokenizer('Today is a nice day and' , return_tensors='tf' ) __UpperCamelCase : Union[str, Any] =tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): __UpperCamelCase : Any =model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ , seed=[7, 0] ) __UpperCamelCase : Tuple =tokenizer.decode(output_ids[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : List[Any] =( 'Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due' ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Optional[Any] =XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) __UpperCamelCase : Optional[Any] ='left' # use different length sentences to test batching __UpperCamelCase : Optional[int] =[ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When', 'Hello, my dog is a little', ] __UpperCamelCase : List[Any] =tokenizer(lowerCamelCase__ , return_tensors='tf' , padding=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =inputs['input_ids'] __UpperCamelCase : Dict =model.generate(input_ids=lowerCamelCase__ , attention_mask=inputs['attention_mask'] , max_new_tokens=12 ) __UpperCamelCase : List[Any] =tokenizer(sentences[0] , return_tensors='tf' ).input_ids __UpperCamelCase : Dict =model.generate(input_ids=lowerCamelCase__ , max_new_tokens=12 ) __UpperCamelCase : Any =tokenizer(sentences[1] , return_tensors='tf' ).input_ids __UpperCamelCase : Optional[Any] =model.generate(input_ids=lowerCamelCase__ , max_new_tokens=12 ) __UpperCamelCase : Optional[int] =tokenizer.batch_decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : int =tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCamelCase__ ) __UpperCamelCase : Any =[ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ' 'a single', 'Hello, my dog is a little bit of a shy one, but he is very friendly', ] self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , [non_padded_sentence, padded_sentence] )
71
1
import copy import random from transformers import CLIPTokenizer class __A ( a ): """simple docstring""" def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) __UpperCamelCase : str ={} def __lowercase ( self , lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Any =super().add_tokens(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) if num_added_tokens == 0: raise ValueError( f'The tokenizer already contains the token {placeholder_token}. Please pass a different' ' `placeholder_token` that is not already in the tokenizer.' ) def __lowercase ( self , lowerCamelCase__ , *lowerCamelCase__ , lowerCamelCase__=1 , **lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] =[] if num_vec_per_token == 1: self.try_adding_tokens(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) output.append(lowerCamelCase__ ) else: __UpperCamelCase : List[str] =[] for i in range(lowerCamelCase__ ): __UpperCamelCase : Dict =placeholder_token + f'_{i}' self.try_adding_tokens(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) output.append(lowerCamelCase__ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( f'The tokenizer already has placeholder token {token} that can get confused with' f' {placeholder_token}keep placeholder tokens independent' ) __UpperCamelCase : List[Any] =output def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=1.0 ): """simple docstring""" if isinstance(lowerCamelCase__ , lowerCamelCase__ ): __UpperCamelCase : Dict =[] for i in range(len(lowerCamelCase__ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=lowerCamelCase__ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: __UpperCamelCase : str =self.token_map[placeholder_token] __UpperCamelCase : Dict =tokens[: 1 + int(len(lowerCamelCase__ ) * prop_tokens_to_load )] if vector_shuffle: __UpperCamelCase : int =copy.copy(lowerCamelCase__ ) random.shuffle(lowerCamelCase__ ) __UpperCamelCase : Dict =text.replace(lowerCamelCase__ , ' '.join(lowerCamelCase__ ) ) return text def __call__( self , lowerCamelCase__ , *lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=1.0 , **lowerCamelCase__ ): """simple docstring""" return super().__call__( self.replace_placeholder_tokens_in_text( lowerCamelCase__ , vector_shuffle=lowerCamelCase__ , prop_tokens_to_load=lowerCamelCase__ ) , *lowerCamelCase__ , **lowerCamelCase__ , ) def __lowercase ( self , lowerCamelCase__ , *lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=1.0 , **lowerCamelCase__ ): """simple docstring""" return super().encode( self.replace_placeholder_tokens_in_text( lowerCamelCase__ , vector_shuffle=lowerCamelCase__ , prop_tokens_to_load=lowerCamelCase__ ) , *lowerCamelCase__ , **lowerCamelCase__ , )
71
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A ( a_ ,a_ ) -> Optional[Any]: # Load checkpoint __UpperCamelCase : int =torch.load(a_ ,map_location='cpu' ) __UpperCamelCase : List[Any] =chkpt['model'] # We have the base model one level deeper than the original XLM repository __UpperCamelCase : str ={} for k, v in state_dict.items(): if "pred_layer" in k: __UpperCamelCase : Optional[Any] =v else: __UpperCamelCase : Optional[Any] =v __UpperCamelCase : List[Any] =chkpt['params'] __UpperCamelCase : str ={n: v for n, v in config.items() if not isinstance(a_ ,(torch.FloatTensor, numpy.ndarray) )} __UpperCamelCase : str =chkpt['dico_word2id'] __UpperCamelCase : Dict ={s + '</w>' if s.find('@@' ) == -1 and i > 13 else s.replace('@@' ,'' ): i for s, i in vocab.items()} # Save pytorch-model __UpperCamelCase : List[Any] =pytorch_dump_folder_path + '/' + WEIGHTS_NAME __UpperCamelCase : Tuple =pytorch_dump_folder_path + '/' + CONFIG_NAME __UpperCamelCase : Any =pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['vocab_file'] print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(a_ ,a_ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(a_ ,indent=2 ) + '\n' ) print(F'Save vocab file to {pytorch_config_dump_path}' ) with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(a_ ,indent=2 ) + '\n' ) if __name__ == "__main__": A_ :str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xlm_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) A_ :List[Any] = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
71
1
A_ :Optional[int] = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' A_ :Union[str, Any] = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] A_ :Optional[Any] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
71
import json import os import shutil import tempfile from unittest import TestCase from transformers import BartTokenizer, BartTokenizerFast, DPRQuestionEncoderTokenizer, DPRQuestionEncoderTokenizerFast from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_tokenizers, require_torch, slow from transformers.utils import is_datasets_available, is_faiss_available, is_torch_available if is_torch_available() and is_datasets_available() and is_faiss_available(): from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.tokenization_rag import RagTokenizer @require_faiss @require_torch class __A ( a ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =tempfile.mkdtemp() __UpperCamelCase : Optional[int] =8 # DPR tok __UpperCamelCase : str =[ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] __UpperCamelCase : Optional[Any] =os.path.join(self.tmpdirname , 'dpr_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , DPR_VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) # BART tok __UpperCamelCase : Optional[int] =[ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] __UpperCamelCase : str =dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) __UpperCamelCase : Optional[int] =['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] __UpperCamelCase : Any ={'unk_token': '<unk>'} __UpperCamelCase : Any =os.path.join(self.tmpdirname , 'bart_tokenizer' ) os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __UpperCamelCase : Any =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['vocab_file'] ) __UpperCamelCase : Dict =os.path.join(lowerCamelCase__ , BART_VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(lowerCamelCase__ ) ) def __lowercase ( self ): """simple docstring""" return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'dpr_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , 'bart_tokenizer' ) ) def __lowercase ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) @require_tokenizers def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =os.path.join(self.tmpdirname , 'rag_tokenizer' ) __UpperCamelCase : Dict =RagConfig(question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() ) __UpperCamelCase : List[Any] =RagTokenizer(question_encoder=self.get_dpr_tokenizer() , generator=self.get_bart_tokenizer() ) rag_config.save_pretrained(lowerCamelCase__ ) rag_tokenizer.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : int =RagTokenizer.from_pretrained(lowerCamelCase__ , config=lowerCamelCase__ ) self.assertIsInstance(new_rag_tokenizer.question_encoder , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.question_encoder.get_vocab() , rag_tokenizer.question_encoder.get_vocab() ) self.assertIsInstance(new_rag_tokenizer.generator , lowerCamelCase__ ) self.assertEqual(new_rag_tokenizer.generator.get_vocab() , rag_tokenizer.generator.get_vocab() ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =RagTokenizer.from_pretrained('facebook/rag-token-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : int =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =RagTokenizer.from_pretrained('facebook/rag-sequence-nq' ) __UpperCamelCase : Union[str, Any] =[ 'who got the first nobel prize in physics', 'when is the next deadpool movie being released', 'which mode is used for short wave broadcast service', 'who is the owner of reading football club', 'when is the next scandal episode coming out', 'when is the last time the philadelphia won the superbowl', 'what is the most current adobe flash player version', 'how many episodes are there in dragon ball z', 'what is the first step in the evolution of the eye', 'where is gall bladder situated in human body', 'what is the main mineral in lithium batteries', 'who is the president of usa right now', 'where do the greasers live in the outsiders', 'panda is a national animal of which country', 'what is the name of manchester united stadium', ] __UpperCamelCase : Any =tokenizer(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
71
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation def A ( a_ ) -> Union[str, Any]: __UpperCamelCase : int =384 __UpperCamelCase : Optional[int] =7 if "tiny" in model_name: __UpperCamelCase : Optional[int] =96 __UpperCamelCase : str =(2, 2, 6, 2) __UpperCamelCase : Union[str, Any] =(3, 6, 12, 24) elif "small" in model_name: __UpperCamelCase : Any =96 __UpperCamelCase : Any =(2, 2, 18, 2) __UpperCamelCase : int =(3, 6, 12, 24) elif "base" in model_name: __UpperCamelCase : Dict =128 __UpperCamelCase : List[str] =(2, 2, 18, 2) __UpperCamelCase : List[Any] =(4, 8, 16, 32) __UpperCamelCase : List[str] =12 __UpperCamelCase : Optional[Any] =512 elif "large" in model_name: __UpperCamelCase : Tuple =192 __UpperCamelCase : Union[str, Any] =(2, 2, 18, 2) __UpperCamelCase : List[Any] =(6, 12, 24, 48) __UpperCamelCase : List[Any] =12 __UpperCamelCase : Union[str, Any] =768 # set label information __UpperCamelCase : Optional[Any] =150 __UpperCamelCase : Any ='huggingface/label-files' __UpperCamelCase : Any ='ade20k-id2label.json' __UpperCamelCase : Tuple =json.load(open(hf_hub_download(a_ ,a_ ,repo_type='dataset' ) ,'r' ) ) __UpperCamelCase : Optional[Any] ={int(a_ ): v for k, v in idalabel.items()} __UpperCamelCase : Optional[Any] ={v: k for k, v in idalabel.items()} __UpperCamelCase : Any =SwinConfig( embed_dim=a_ ,depths=a_ ,num_heads=a_ ,window_size=a_ ,out_features=['stage1', 'stage2', 'stage3', 'stage4'] ,) __UpperCamelCase : str =UperNetConfig( backbone_config=a_ ,auxiliary_in_channels=a_ ,num_labels=a_ ,idalabel=a_ ,labelaid=a_ ,) return config def A ( a_ ) -> int: __UpperCamelCase : Any =[] # fmt: off # stem rename_keys.append(('backbone.patch_embed.projection.weight', 'backbone.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('backbone.patch_embed.projection.bias', 'backbone.embeddings.patch_embeddings.projection.bias') ) rename_keys.append(('backbone.patch_embed.norm.weight', 'backbone.embeddings.norm.weight') ) rename_keys.append(('backbone.patch_embed.norm.bias', 'backbone.embeddings.norm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm1.weight', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm1.bias', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table', F'backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index', F'backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight', F'backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias', F'backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm2.weight', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm2.bias', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight', F'backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias', F'backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight', F'backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias', F'backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias') ) if i < 3: rename_keys.append((F'backbone.stages.{i}.downsample.reduction.weight', F'backbone.encoder.layers.{i}.downsample.reduction.weight') ) rename_keys.append((F'backbone.stages.{i}.downsample.norm.weight', F'backbone.encoder.layers.{i}.downsample.norm.weight') ) rename_keys.append((F'backbone.stages.{i}.downsample.norm.bias', F'backbone.encoder.layers.{i}.downsample.norm.bias') ) rename_keys.append((F'backbone.norm{i}.weight', F'backbone.hidden_states_norms.stage{i+1}.weight') ) rename_keys.append((F'backbone.norm{i}.bias', F'backbone.hidden_states_norms.stage{i+1}.bias') ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def A ( a_ ,a_ ,a_ ) -> List[Any]: __UpperCamelCase : Any =dct.pop(a_ ) __UpperCamelCase : Union[str, Any] =val def A ( a_ ,a_ ) -> Union[str, Any]: __UpperCamelCase : Optional[Any] =[int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __UpperCamelCase : Any =num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __UpperCamelCase : str =state_dict.pop(F'backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight' ) __UpperCamelCase : str =state_dict.pop(F'backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict __UpperCamelCase : Any =in_proj_weight[:dim, :] __UpperCamelCase : List[Any] =in_proj_bias[: dim] __UpperCamelCase : str =in_proj_weight[ dim : dim * 2, : ] __UpperCamelCase : List[str] =in_proj_bias[ dim : dim * 2 ] __UpperCamelCase : Optional[Any] =in_proj_weight[ -dim :, : ] __UpperCamelCase : Optional[int] =in_proj_bias[-dim :] # fmt: on def A ( a_ ) -> Optional[int]: __UpperCamelCase , __UpperCamelCase : Any =x.shape __UpperCamelCase : Union[str, Any] =x.reshape(a_ ,4 ,in_channel // 4 ) __UpperCamelCase : List[Any] =x[:, [0, 2, 1, 3], :].transpose(1 ,2 ).reshape(a_ ,a_ ) return x def A ( a_ ) -> List[str]: __UpperCamelCase , __UpperCamelCase : str =x.shape __UpperCamelCase : str =x.reshape(a_ ,in_channel // 4 ,4 ) __UpperCamelCase : List[Any] =x[:, :, [0, 2, 1, 3]].transpose(1 ,2 ).reshape(a_ ,a_ ) return x def A ( a_ ) -> Any: __UpperCamelCase : Dict =x.shape[0] __UpperCamelCase : Tuple =x.reshape(4 ,in_channel // 4 ) __UpperCamelCase : Optional[Any] =x[[0, 2, 1, 3], :].transpose(0 ,1 ).reshape(a_ ) return x def A ( a_ ) -> Dict: __UpperCamelCase : List[Any] =x.shape[0] __UpperCamelCase : List[str] =x.reshape(in_channel // 4 ,4 ) __UpperCamelCase : Tuple =x[:, [0, 2, 1, 3]].transpose(0 ,1 ).reshape(a_ ) return x def A ( a_ ,a_ ,a_ ) -> Optional[Any]: __UpperCamelCase : int ={ 'upernet-swin-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth', 'upernet-swin-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth', 'upernet-swin-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth', 'upernet-swin-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth', } __UpperCamelCase : int =model_name_to_url[model_name] __UpperCamelCase : List[str] =torch.hub.load_state_dict_from_url(a_ ,map_location='cpu' ,file_name=a_ )[ 'state_dict' ] for name, param in state_dict.items(): print(a_ ,param.shape ) __UpperCamelCase : List[str] =get_upernet_config(a_ ) __UpperCamelCase : str =UperNetForSemanticSegmentation(a_ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): __UpperCamelCase : Optional[int] =state_dict.pop(a_ ) if "bn" in key: __UpperCamelCase : Union[str, Any] =key.replace('bn' ,'batch_norm' ) __UpperCamelCase : Union[str, Any] =val # rename keys __UpperCamelCase : int =create_rename_keys(a_ ) for src, dest in rename_keys: rename_key(a_ ,a_ ,a_ ) read_in_q_k_v(a_ ,config.backbone_config ) # fix downsample parameters for key, value in state_dict.items(): if "downsample" in key: if "reduction" in key: __UpperCamelCase : str =reverse_correct_unfold_reduction_order(a_ ) if "norm" in key: __UpperCamelCase : str =reverse_correct_unfold_norm_order(a_ ) model.load_state_dict(a_ ) # verify on image __UpperCamelCase : Any ='https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' __UpperCamelCase : List[Any] =Image.open(requests.get(a_ ,stream=a_ ).raw ).convert('RGB' ) __UpperCamelCase : Union[str, Any] =SegformerImageProcessor() __UpperCamelCase : Optional[Any] =processor(a_ ,return_tensors='pt' ).pixel_values with torch.no_grad(): __UpperCamelCase : Optional[int] =model(a_ ) __UpperCamelCase : int =outputs.logits print(logits.shape ) print('First values of logits:' ,logits[0, 0, :3, :3] ) # assert values if model_name == "upernet-swin-tiny": __UpperCamelCase : List[Any] =torch.tensor( [[-7.5_958, -7.5_958, -7.4_302], [-7.5_958, -7.5_958, -7.4_302], [-7.4_797, -7.4_797, -7.3_068]] ) elif model_name == "upernet-swin-small": __UpperCamelCase : Union[str, Any] =torch.tensor( [[-7.1_921, -7.1_921, -6.9_532], [-7.1_921, -7.1_921, -6.9_532], [-7.0_908, -7.0_908, -6.8_534]] ) elif model_name == "upernet-swin-base": __UpperCamelCase : Tuple =torch.tensor( [[-6.5_851, -6.5_851, -6.4_330], [-6.5_851, -6.5_851, -6.4_330], [-6.4_763, -6.4_763, -6.3_254]] ) elif model_name == "upernet-swin-large": __UpperCamelCase : Optional[Any] =torch.tensor( [[-7.5_297, -7.5_297, -7.3_802], [-7.5_297, -7.5_297, -7.3_802], [-7.4_044, -7.4_044, -7.2_586]] ) print('Logits:' ,outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] ,a_ ,atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(a_ ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(a_ ) if push_to_hub: print(F'Pushing model and processor for {model_name} to hub' ) model.push_to_hub(F'openmmlab/{model_name}' ) processor.push_to_hub(F'openmmlab/{model_name}' ) if __name__ == "__main__": A_ :Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''upernet-swin-tiny''', type=str, choices=[f"upernet-swin-{size}" for size in ['''tiny''', '''small''', '''base''', '''large''']], help='''Name of the Swin + UperNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ :Optional[Any] = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
71
A_ :Optional[int] = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' A_ :Union[str, Any] = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] A_ :Optional[Any] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
71
1
import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __A ( a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Tuple =KandinskyVaaControlnetPipeline UpperCamelCase__ : Dict =["""image_embeds""", """negative_image_embeds""", """hint"""] UpperCamelCase__ : int =["""image_embeds""", """negative_image_embeds""", """hint"""] UpperCamelCase__ : List[str] =[ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] UpperCamelCase__ : Tuple =False @property def __lowercase ( self ): """simple docstring""" return 32 @property def __lowercase ( self ): """simple docstring""" return 32 @property def __lowercase ( self ): """simple docstring""" return self.time_input_dim @property def __lowercase ( self ): """simple docstring""" return self.time_input_dim * 4 @property def __lowercase ( self ): """simple docstring""" return 100 @property def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Tuple ={ 'in_channels': 8, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image_hint', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } __UpperCamelCase : Optional[Any] =UNetaDConditionModel(**lowerCamelCase__ ) return model @property def __lowercase ( self ): """simple docstring""" return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : List[Any] =VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =self.dummy_unet __UpperCamelCase : List[Any] =self.dummy_movq __UpperCamelCase : int =DDIMScheduler( num_train_timesteps=1000 , beta_schedule='linear' , beta_start=0.00_085 , beta_end=0.012 , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , steps_offset=1 , prediction_type='epsilon' , thresholding=lowerCamelCase__ , ) __UpperCamelCase : Optional[int] ={ 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : Any =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : List[str] =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( lowerCamelCase__ ) # create hint __UpperCamelCase : List[str] =floats_tensor((1, 3, 64, 64) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Optional[int] =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : List[Any] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Tuple ={ 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'hint': hint, 'generator': generator, 'height': 64, 'width': 64, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str ='cpu' __UpperCamelCase : Tuple =self.get_dummy_components() __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =pipe(**self.get_dummy_inputs(lowerCamelCase__ ) ) __UpperCamelCase : Union[str, Any] =output.images __UpperCamelCase : Optional[Any] =pipe( **self.get_dummy_inputs(lowerCamelCase__ ) , return_dict=lowerCamelCase__ , )[0] __UpperCamelCase : Tuple =image[0, -3:, -3:, -1] __UpperCamelCase : List[str] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __UpperCamelCase : List[Any] =np.array( [0.6_959_826, 0.868_279, 0.7_558_092, 0.68_769_467, 0.85_805_804, 0.65_977_496, 0.44_885_302, 0.5_959_111, 0.4_251_595] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy' ) __UpperCamelCase : Union[str, Any] =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/hint_image_cat.png' ) __UpperCamelCase : Any =torch.from_numpy(np.array(lowerCamelCase__ ) ).float() / 255.0 __UpperCamelCase : int =hint.permute(2 , 0 , 1 ).unsqueeze(0 ) __UpperCamelCase : Dict =KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(lowerCamelCase__ ) __UpperCamelCase : Dict =KandinskyVaaControlnetPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-controlnet-depth' , torch_dtype=torch.floataa ) __UpperCamelCase : List[Any] =pipeline.to(lowerCamelCase__ ) pipeline.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple ='A robot, 4k photo' __UpperCamelCase : Optional[int] =torch.Generator(device='cuda' ).manual_seed(0 ) __UpperCamelCase , __UpperCamelCase : List[str] =pipe_prior( lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=5 , negative_prompt='' , ).to_tuple() __UpperCamelCase : Any =torch.Generator(device='cuda' ).manual_seed(0 ) __UpperCamelCase : str =pipeline( image_embeds=lowerCamelCase__ , negative_image_embeds=lowerCamelCase__ , hint=lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=100 , output_type='np' , ) __UpperCamelCase : Optional[int] =output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(lowerCamelCase__ , lowerCamelCase__ )
71
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration A_ :Optional[Any] = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def A ( a_ ) -> List[Any]: __UpperCamelCase : Any =['layers', 'blocks'] for k in ignore_keys: state_dict.pop(a_ ,a_ ) A_ :int = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def A ( a_ ) -> Union[str, Any]: __UpperCamelCase : str =list(s_dict.keys() ) for key in keys: __UpperCamelCase : str =key for k, v in WHISPER_MAPPING.items(): if k in key: __UpperCamelCase : Optional[Any] =new_key.replace(a_ ,a_ ) print(F'{key} -> {new_key}' ) __UpperCamelCase : Dict =s_dict.pop(a_ ) return s_dict def A ( a_ ) -> Optional[Any]: __UpperCamelCase , __UpperCamelCase : Tuple =emb.weight.shape __UpperCamelCase : Tuple =nn.Linear(a_ ,a_ ,bias=a_ ) __UpperCamelCase : List[Any] =emb.weight.data return lin_layer def A ( a_ ,a_ ) -> bytes: os.makedirs(a_ ,exist_ok=a_ ) __UpperCamelCase : Optional[int] =os.path.basename(a_ ) __UpperCamelCase : Union[str, Any] =url.split('/' )[-2] __UpperCamelCase : Union[str, Any] =os.path.join(a_ ,a_ ) if os.path.exists(a_ ) and not os.path.isfile(a_ ): raise RuntimeError(F'{download_target} exists and is not a regular file' ) if os.path.isfile(a_ ): __UpperCamelCase : str =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(F'{download_target} exists, but the SHA256 checksum does not match; re-downloading the file' ) with urllib.request.urlopen(a_ ) as source, open(a_ ,'wb' ) as output: with tqdm( total=int(source.info().get('Content-Length' ) ) ,ncols=80 ,unit='iB' ,unit_scale=a_ ,unit_divisor=1_024 ) as loop: while True: __UpperCamelCase : Optional[Any] =source.read(8_192 ) if not buffer: break output.write(a_ ) loop.update(len(a_ ) ) __UpperCamelCase : List[Any] =open(a_ ,'rb' ).read() if hashlib.shaaaa(a_ ).hexdigest() != expected_shaaaa: raise RuntimeError( 'Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.' ) return model_bytes def A ( a_ ,a_ ) -> Optional[Any]: if ".pt" not in checkpoint_path: __UpperCamelCase : int =_download(_MODELS[checkpoint_path] ) else: __UpperCamelCase : List[str] =torch.load(a_ ,map_location='cpu' ) __UpperCamelCase : Union[str, Any] =original_checkpoint['dims'] __UpperCamelCase : List[Any] =original_checkpoint['model_state_dict'] __UpperCamelCase : Dict =state_dict['decoder.token_embedding.weight'] remove_ignore_keys_(a_ ) rename_keys(a_ ) __UpperCamelCase : List[str] =True __UpperCamelCase : str =state_dict['decoder.layers.0.fc1.weight'].shape[0] __UpperCamelCase : Optional[int] =WhisperConfig( vocab_size=dimensions['n_vocab'] ,encoder_ffn_dim=a_ ,decoder_ffn_dim=a_ ,num_mel_bins=dimensions['n_mels'] ,d_model=dimensions['n_audio_state'] ,max_target_positions=dimensions['n_text_ctx'] ,encoder_layers=dimensions['n_audio_layer'] ,encoder_attention_heads=dimensions['n_audio_head'] ,decoder_layers=dimensions['n_text_layer'] ,decoder_attention_heads=dimensions['n_text_state'] ,max_source_positions=dimensions['n_audio_ctx'] ,) __UpperCamelCase : List[str] =WhisperForConditionalGeneration(a_ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =model.model.load_state_dict(a_ ,strict=a_ ) if len(a_ ) > 0 and not set(a_ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( 'Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,' F' but all the following weights are missing {missing}' ) if tie_embeds: __UpperCamelCase : Optional[int] =make_linear_from_emb(model.model.decoder.embed_tokens ) else: __UpperCamelCase : List[str] =proj_out_weights model.save_pretrained(a_ ) if __name__ == "__main__": A_ :List[Any] = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') A_ :List[Any] = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
71
1
import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py A_ :Tuple = '''src/transformers''' A_ :str = '''docs/source/en/tasks''' def A ( a_ ,a_ ,a_ ) -> Optional[int]: with open(a_ ,'r' ,encoding='utf-8' ,newline='\n' ) as f: __UpperCamelCase : Optional[int] =f.readlines() # Find the start prompt. __UpperCamelCase : Optional[Any] =0 while not lines[start_index].startswith(a_ ): start_index += 1 start_index += 1 __UpperCamelCase : Any =start_index while not lines[end_index].startswith(a_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. A_ :int = direct_transformers_import(TRANSFORMERS_PATH) A_ :Union[str, Any] = { '''asr.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, '''audio_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, '''language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, '''image_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, '''masked_language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, '''multiple_choice.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, '''object_detection.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, '''question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, '''semantic_segmentation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, '''sequence_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, '''summarization.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''token_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, '''translation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''video_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, '''document_question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, '''monocular_depth_estimation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). A_ :Optional[Any] = { '''summarization.md''': ('''nllb''',), '''translation.md''': ('''nllb''',), } def A ( a_ ) -> Tuple: __UpperCamelCase : int =TASK_GUIDE_TO_MODELS[task_guide] __UpperCamelCase : Dict =SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(a_ ,set() ) __UpperCamelCase : List[Any] ={ code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F'[{name}](../model_doc/{code})' for code, name in model_names.items()] ) + "\n" def A ( a_ ,a_=False ) -> Optional[int]: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase : Tuple =_find_text_in_file( filename=os.path.join(a_ ,a_ ) ,start_prompt='<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->' ,end_prompt='<!--End of the generated tip-->' ,) __UpperCamelCase : Optional[Any] =get_model_list_for_task(a_ ) if current_list != new_list: if overwrite: with open(os.path.join(a_ ,a_ ) ,'w' ,encoding='utf-8' ,newline='\n' ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F'The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`' ' to fix this.' ) if __name__ == "__main__": A_ :List[Any] = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ :int = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
71
import os from datetime import datetime as dt from github import Github A_ :str = [ '''good first issue''', '''feature request''', '''wip''', ] def A ( ) -> Any: __UpperCamelCase : Any =Github(os.environ['GITHUB_TOKEN'] ) __UpperCamelCase : Union[str, Any] =g.get_repo('huggingface/accelerate' ) __UpperCamelCase : Tuple =repo.get_issues(state='open' ) for issue in open_issues: __UpperCamelCase : List[Any] =sorted([comment for comment in issue.get_comments()] ,key=lambda a_ : i.created_at ,reverse=a_ ) __UpperCamelCase : str =comments[0] if len(a_ ) > 0 else None __UpperCamelCase : Any =dt.utcnow() __UpperCamelCase : List[str] =(current_time - issue.updated_at).days __UpperCamelCase : Union[str, Any] =(current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state='closed' ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
71
1
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def A ( a_=None ,a_=None ) -> Dict: return field(default_factory=lambda: default ,metadata=a_ ) @dataclass class __A : """simple docstring""" UpperCamelCase__ : str =field( metadata={"""help""": """The csv file to plot."""} , ) UpperCamelCase__ : bool =field( default=a , metadata={"""help""": """Whether to plot along batch size or sequence length. Defaults to sequence length."""} , ) UpperCamelCase__ : bool =field( default=a , metadata={"""help""": """Whether the csv file has time results or memory results. Defaults to memory results."""} , ) UpperCamelCase__ : bool =field( default=a , metadata={"""help""": """Disable logarithmic scale when plotting"""} , ) UpperCamelCase__ : bool =field( default=a , metadata={ """help""": """Whether the csv file has training results or inference results. Defaults to inference results.""" } , ) UpperCamelCase__ : Optional[str] =field( default=a , metadata={"""help""": """Filename under which the plot will be saved. If unused no plot is saved."""} , ) UpperCamelCase__ : Optional[List[str]] =list_field( default=a , metadata={"""help""": """List of model names that are used instead of the ones in the csv file."""} ) def A ( a_ ) -> List[str]: try: int(a_ ) return True except ValueError: return False def A ( a_ ) -> List[Any]: try: float(a_ ) return True except ValueError: return False class __A : """simple docstring""" def __init__( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Union[str, Any] =args __UpperCamelCase : List[str] =defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline='' ) as csv_file: __UpperCamelCase : Optional[Any] =csv.DictReader(lowerCamelCase__ ) for row in reader: __UpperCamelCase : Tuple =row['model'] self.result_dict[model_name]["bsz"].append(int(row['batch_size'] ) ) self.result_dict[model_name]["seq_len"].append(int(row['sequence_length'] ) ) if can_convert_to_int(row['result'] ): # value is not None __UpperCamelCase : Any =int(row['result'] ) elif can_convert_to_float(row['result'] ): # value is not None __UpperCamelCase : str =float(row['result'] ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : List[Any] =plt.subplots() __UpperCamelCase : str ='Time usage' if self.args.is_time else 'Memory usage' __UpperCamelCase : Tuple =title_str + ' for training' if self.args.is_train else title_str + ' for inference' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('log' ) ax.set_yscale('log' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): __UpperCamelCase : Union[str, Any] =sorted(set(self.result_dict[model_name]['bsz'] ) ) __UpperCamelCase : Dict =sorted(set(self.result_dict[model_name]['seq_len'] ) ) __UpperCamelCase : Any =self.result_dict[model_name]['result'] ((__UpperCamelCase) , (__UpperCamelCase)) : Dict =( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) __UpperCamelCase : Any =( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: __UpperCamelCase : Union[str, Any] =np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=lowerCamelCase__ , ) else: __UpperCamelCase : Union[str, Any] =np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((__UpperCamelCase) , (__UpperCamelCase)) : Optional[int] =( ('batch_size', 'len') if self.args.plot_along_batch else ('in #tokens', 'bsz') ) __UpperCamelCase : Any =np.asarray(lowerCamelCase__ , lowerCamelCase__ )[: len(lowerCamelCase__ )] plt.scatter( lowerCamelCase__ , lowerCamelCase__ , label=f'{label_model_name} - {inner_loop_label}: {inner_loop_value}' ) plt.plot(lowerCamelCase__ , lowerCamelCase__ , '--' ) title_str += f' {label_model_name} vs.' __UpperCamelCase : int =title_str[:-4] __UpperCamelCase : Tuple ='Time in s' if self.args.is_time else 'Memory in MB' # plot plt.title(lowerCamelCase__ ) plt.xlabel(lowerCamelCase__ ) plt.ylabel(lowerCamelCase__ ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def A ( ) -> List[Any]: __UpperCamelCase : Optional[int] =HfArgumentParser(a_ ) __UpperCamelCase : Any =parser.parse_args_into_dataclasses()[0] __UpperCamelCase : Optional[int] =Plot(args=a_ ) plot.plot() if __name__ == "__main__": main()
71
import re def A ( a_ ) -> bool: __UpperCamelCase : Any =re.compile( r'^(?:0|94|\+94|0{2}94)' r'7(0|1|2|4|5|6|7|8)' r'(-| |)' r'\d{7}$' ) return bool(re.search(a_ ,a_ ) ) if __name__ == "__main__": A_ :List[str] = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
71
1
# Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def A ( a_ ) -> Optional[Any]: return 1 / (1 + np.exp(-z )) def A ( a_ ,a_ ) -> int: return (-y * np.log(a_ ) - (1 - y) * np.log(1 - h )).mean() def A ( a_ ,a_ ,a_ ) -> Tuple: __UpperCamelCase : Optional[int] =np.dot(a_ ,a_ ) return np.sum(y * scores - np.log(1 + np.exp(a_ ) ) ) def A ( a_ ,a_ ,a_ ,a_=70_000 ) -> str: __UpperCamelCase : Union[str, Any] =np.zeros(x.shape[1] ) for iterations in range(a_ ): __UpperCamelCase : str =np.dot(a_ ,a_ ) __UpperCamelCase : Any =sigmoid_function(a_ ) __UpperCamelCase : Optional[Any] =np.dot(x.T ,h - y ) / y.size __UpperCamelCase : int =theta - alpha * gradient # updating the weights __UpperCamelCase : Dict =np.dot(a_ ,a_ ) __UpperCamelCase : Optional[Any] =sigmoid_function(a_ ) __UpperCamelCase : List[Any] =cost_function(a_ ,a_ ) if iterations % 100 == 0: print(F'loss: {j} \t' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": A_ :Optional[Any] = datasets.load_iris() A_ :Any = iris.data[:, :2] A_ :Optional[Any] = (iris.target != 0) * 1 A_ :Tuple = 0.1 A_ :Any = logistic_reg(alpha, x, y, max_iterations=70000) print('''theta: ''', theta) # printing the theta i.e our weights vector def A ( a_ ) -> str: return sigmoid_function( np.dot(a_ ,a_ ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((A_) ,(A_)) :Union[str, Any] = (x[:, 0].min(), x[:, 0].max()) ((A_) ,(A_)) :Optional[Any] = (x[:, 1].min(), x[:, 1].max()) ((A_) ,(A_)) :Tuple = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) A_ :List[str] = np.c_[xxa.ravel(), xxa.ravel()] A_ :Optional[int] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
71
A_ :str = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
71
1
from __future__ import annotations import typing from collections import Counter def A ( a_ ) -> typing.Counter[int]: __UpperCamelCase : typing.Counter[int] =Counter() for base in range(1 ,max_perimeter + 1 ): for perpendicular in range(a_ ,max_perimeter + 1 ): __UpperCamelCase : List[Any] =(base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(a_ ): __UpperCamelCase : Tuple =int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def A ( a_ = 1_000 ) -> int: __UpperCamelCase : Dict =pythagorean_triple(a_ ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(f"Perimeter {solution()} has maximum solutions")
71
A_ :Union[str, Any] = { 0: '''0''', 1: '''1''', 2: '''2''', 3: '''3''', 4: '''4''', 5: '''5''', 6: '''6''', 7: '''7''', 8: '''8''', 9: '''9''', 10: '''a''', 11: '''b''', 12: '''c''', 13: '''d''', 14: '''e''', 15: '''f''', } def A ( a_ ) -> str: assert type(a_ ) in (int, float) and decimal == int(a_ ) __UpperCamelCase : Union[str, Any] =int(a_ ) __UpperCamelCase : List[str] ='' __UpperCamelCase : Optional[Any] =False if decimal < 0: __UpperCamelCase : Tuple =True decimal *= -1 while decimal > 0: __UpperCamelCase , __UpperCamelCase : Optional[Any] =divmod(a_ ,16 ) __UpperCamelCase : Tuple =values[remainder] + hexadecimal __UpperCamelCase : Dict ='0x' + hexadecimal if negative: __UpperCamelCase : int ='-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
71
1
from __future__ import annotations from collections.abc import Iterator class __A : """simple docstring""" def __init__( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Union[str, Any] =value __UpperCamelCase : Node | None =None __UpperCamelCase : Node | None =None class __A : """simple docstring""" def __init__( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : List[str] =tree def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ): """simple docstring""" yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
71
import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness A_ :List[str] = '''\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } ''' A_ :Any = '''\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). ''' A_ :Tuple = ''' Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric("code_eval") >>> test_cases = ["assert add(2,3)==5"] >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {\'pass@1\': 0.5, \'pass@2\': 1.0} ''' A_ :List[str] = ''' ################################################################################ !!!WARNING!!! ################################################################################ The "code_eval" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this with: >>> import os >>> os.environ["HF_ALLOW_CODE_EVAL"] = "1" ################################################################################\ ''' A_ :Tuple = '''The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): """simple docstring""" def __lowercase ( self ): """simple docstring""" return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' ) ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/openai/human-eval' , codebase_urls=['https://github.com/openai/human-eval'] , reference_urls=['https://github.com/openai/human-eval'] , license=_LICENSE , ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=[1, 10, 100] , lowerCamelCase__=4 , lowerCamelCase__=3.0 ): """simple docstring""" if os.getenv('HF_ALLOW_CODE_EVAL' , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError('This metric is currently not supported on Windows.' ) with ThreadPoolExecutor(max_workers=lowerCamelCase__ ) as executor: __UpperCamelCase : List[str] =[] __UpperCamelCase : Any =Counter() __UpperCamelCase : List[Any] =0 __UpperCamelCase : int =defaultdict(lowerCamelCase__ ) for task_id, (candidates, test_case) in enumerate(zip(lowerCamelCase__ , lowerCamelCase__ ) ): for candidate in candidates: __UpperCamelCase : str =candidate + '\n' + test_case __UpperCamelCase : Any =(test_program, timeout, task_id, completion_id[task_id]) __UpperCamelCase : Optional[Any] =executor.submit(lowerCamelCase__ , *lowerCamelCase__ ) futures.append(lowerCamelCase__ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(lowerCamelCase__ ): __UpperCamelCase : str =future.result() results[result["task_id"]].append((result['completion_id'], result) ) __UpperCamelCase , __UpperCamelCase : int =[], [] for result in results.values(): result.sort() __UpperCamelCase : str =[r[1]['passed'] for r in result] total.append(len(lowerCamelCase__ ) ) correct.append(sum(lowerCamelCase__ ) ) __UpperCamelCase : Optional[int] =np.array(lowerCamelCase__ ) __UpperCamelCase : List[str] =np.array(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =k __UpperCamelCase : List[Any] ={f'pass@{k}': estimate_pass_at_k(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def A ( a_ ,a_ ,a_ ) -> Optional[int]: def estimator(a_ ,a_ ,a_ ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 ,n + 1 ) ) if isinstance(a_ ,a_ ): __UpperCamelCase : Optional[int] =itertools.repeat(a_ ,len(a_ ) ) else: assert len(a_ ) == len(a_ ) __UpperCamelCase : List[Any] =iter(a_ ) return np.array([estimator(int(a_ ) ,int(a_ ) ,a_ ) for n, c in zip(a_ ,a_ )] )
71
1
import argparse from collections import defaultdict import yaml A_ :Tuple = '''docs/source/en/_toctree.yml''' def A ( a_ ) -> Tuple: __UpperCamelCase : Tuple =defaultdict(a_ ) for doc in model_doc: counts[doc["local"]] += 1 __UpperCamelCase : int =[key for key, value in counts.items() if value > 1] __UpperCamelCase : Any =[] for duplicate_key in duplicates: __UpperCamelCase : Union[str, Any] =list({doc['title'] for doc in model_doc if doc['local'] == duplicate_key} ) if len(a_ ) > 1: raise ValueError( F'{duplicate_key} is present several times in the documentation table of content at ' '`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ' 'others.' ) # Only add this once new_doc.append({'local': duplicate_key, 'title': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['local']] == 1] ) # Sort return sorted(a_ ,key=lambda a_ : s["title"].lower() ) def A ( a_=False ) -> Union[str, Any]: with open(a_ ,encoding='utf-8' ) as f: __UpperCamelCase : Any =yaml.safe_load(f.read() ) # Get to the API doc __UpperCamelCase : List[str] =0 while content[api_idx]["title"] != "API": api_idx += 1 __UpperCamelCase : Any =content[api_idx]['sections'] # Then to the model doc __UpperCamelCase : List[str] =0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 __UpperCamelCase : str =api_doc[model_idx]['sections'] __UpperCamelCase : int =[(idx, section) for idx, section in enumerate(a_ ) if 'sections' in section] __UpperCamelCase : Any =False for idx, modality_doc in modalities_docs: __UpperCamelCase : Optional[Any] =modality_doc['sections'] __UpperCamelCase : str =clean_model_doc_toc(a_ ) if old_modality_doc != new_modality_doc: __UpperCamelCase : str =True if overwrite: __UpperCamelCase : List[Any] =new_modality_doc if diff: if overwrite: __UpperCamelCase : Union[str, Any] =model_doc __UpperCamelCase : int =api_doc with open(a_ ,'w' ,encoding='utf-8' ) as f: f.write(yaml.dump(a_ ,allow_unicode=a_ ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) if __name__ == "__main__": A_ :Dict = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ :Any = parser.parse_args() check_model_doc(args.fix_and_overwrite)
71
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __A ( a , a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Optional[Any] =StableDiffusionDiffEditPipeline UpperCamelCase__ : str =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""height""", """width""", """image"""} | {"""image_latents"""} UpperCamelCase__ : Optional[Any] =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"""image"""} | {"""image_latents"""} UpperCamelCase__ : Dict =frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess UpperCamelCase__ : Any =frozenset([] ) def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Dict =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , ) __UpperCamelCase : List[str] =DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) __UpperCamelCase : Union[str, Any] =DDIMInverseScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_zero=lowerCamelCase__ , ) torch.manual_seed(0 ) __UpperCamelCase : Optional[int] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __UpperCamelCase : Tuple =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) __UpperCamelCase : Any =CLIPTextModel(lowerCamelCase__ ) __UpperCamelCase : int =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __UpperCamelCase : Union[str, Any] ={ 'unet': unet, 'scheduler': scheduler, 'inverse_scheduler': inverse_scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : int =floats_tensor((1, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Optional[int] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Dict ={ 'prompt': 'a dog and a newt', 'mask_image': mask, 'image_latents': latents, 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : Tuple =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : int =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : Optional[Any] =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : List[Any] =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Any =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'source_prompt': 'a cat and a frog', 'target_prompt': 'a dog and a newt', 'generator': generator, 'num_inference_steps': 2, 'num_maps_per_mask': 2, 'mask_encode_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : str =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : int =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : int =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'prompt': 'a cat and a frog', 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'decode_latents': True, 'output_type': 'numpy', } return inputs def __lowercase ( self ): """simple docstring""" if not hasattr(self.pipeline_class , '_optional_components' ): return __UpperCamelCase : Optional[Any] =self.get_dummy_components() __UpperCamelCase : List[str] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) __UpperCamelCase : Union[str, Any] =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe(**lowerCamelCase__ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : Tuple =self.pipeline_class.from_pretrained(lowerCamelCase__ ) pipe_loaded.to(lowerCamelCase__ ) pipe_loaded.set_progress_bar_config(disable=lowerCamelCase__ ) for optional_component in pipe._optional_components: self.assertTrue( getattr(lowerCamelCase__ , lowerCamelCase__ ) is None , f'`{optional_component}` did not stay set to None after loading.' , ) __UpperCamelCase : str =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe_loaded(**lowerCamelCase__ )[0] __UpperCamelCase : Tuple =np.abs(output - output_loaded ).max() self.assertLess(lowerCamelCase__ , 1E-4 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : int =self.get_dummy_mask_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe.generate_mask(**lowerCamelCase__ ) __UpperCamelCase : int =mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) __UpperCamelCase : Tuple =np.array([0] * 9 ) __UpperCamelCase : str =np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Optional[Any] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Dict =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : Optional[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : int =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def __lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=5E-3 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ='cpu' __UpperCamelCase : int =self.get_dummy_components() __UpperCamelCase : str ={'beta_start': 0.00_085, 'beta_end': 0.012, 'beta_schedule': 'scaled_linear'} __UpperCamelCase : str =DPMSolverMultistepScheduler(**lowerCamelCase__ ) __UpperCamelCase : Dict =DPMSolverMultistepInverseScheduler(**lowerCamelCase__ ) __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : str =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : List[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : Optional[Any] =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) @require_torch_gpu @slow class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def __lowercase ( cls ): """simple docstring""" __UpperCamelCase : Optional[int] =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png' ) __UpperCamelCase : Union[str, Any] =raw_image.convert('RGB' ).resize((768, 768) ) __UpperCamelCase : List[Any] =raw_image def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =torch.manual_seed(0 ) __UpperCamelCase : Dict =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : List[str] =DDIMScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : List[str] =DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : List[str] ='a bowl of fruit' __UpperCamelCase : Dict ='a bowl of pears' __UpperCamelCase : Tuple =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : int =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ ).latents __UpperCamelCase : Dict =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , output_type='numpy' , ).images[0] __UpperCamelCase : str =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =torch.manual_seed(0 ) __UpperCamelCase : List[Any] =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : Optional[Any] =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : Optional[int] =DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] ='a bowl of fruit' __UpperCamelCase : int ='a bowl of pears' __UpperCamelCase : str =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : List[str] =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ , num_inference_steps=25 , ).latents __UpperCamelCase : List[str] =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , num_inference_steps=25 , output_type='numpy' , ).images[0] __UpperCamelCase : Tuple =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1
71
1
from torch import nn def A ( a_ ) -> Any: if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'Unsupported activation function: {act_fn}' )
71
import random from .binary_exp_mod import bin_exp_mod def A ( a_ ,a_=1_000 ) -> Optional[Any]: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd __UpperCamelCase : List[Any] =n - 1 __UpperCamelCase : Dict =0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) __UpperCamelCase : Optional[Any] =0 while count < prec: __UpperCamelCase : Dict =random.randint(2 ,n - 1 ) __UpperCamelCase : Optional[Any] =bin_exp_mod(a_ ,a_ ,a_ ) if b != 1: __UpperCamelCase : List[str] =True for _ in range(a_ ): if b == n - 1: __UpperCamelCase : Tuple =False break __UpperCamelCase : Dict =b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": A_ :str = abs(int(input('''Enter bound : ''').strip())) print('''Here\'s the list of primes:''') print(''', '''.join(str(i) for i in range(n + 1) if is_prime_big(i)))
71
1
from bisect import bisect from itertools import accumulate def A ( a_ ,a_ ,a_ ,a_ ) -> Optional[Any]: __UpperCamelCase : str =sorted(zip(a_ ,a_ ) ,key=lambda a_ : x[0] / x[1] ,reverse=a_ ) __UpperCamelCase , __UpperCamelCase : int =[i[0] for i in r], [i[1] for i in r] __UpperCamelCase : List[str] =list(accumulate(a_ ) ) __UpperCamelCase : str =bisect(a_ ,a_ ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
71
from torch import nn class __A ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" super().__init__() __UpperCamelCase : Dict =class_size __UpperCamelCase : Any =embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) __UpperCamelCase : Any =nn.Linear(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : List[Any] =self.mlp(lowerCamelCase__ ) return logits
71
1
def A ( a_ ) -> float: return 10 - x * x def A ( a_ ,a_ ) -> float: # Bolzano theory in order to find if there is a root between a and b if equation(a_ ) * equation(a_ ) >= 0: raise ValueError('Wrong space!' ) __UpperCamelCase : Optional[Any] =a while (b - a) >= 0.01: # Find middle point __UpperCamelCase : int =(a + b) / 2 # Check if middle point is root if equation(a_ ) == 0.0: break # Decide the side to repeat the steps if equation(a_ ) * equation(a_ ) < 0: __UpperCamelCase : List[Any] =c else: __UpperCamelCase : Tuple =c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
71
def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square(a_ ,a_ ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __UpperCamelCase : Optional[int] =update_area_of_max_square(a_ ,col + 1 ) __UpperCamelCase : List[str] =update_area_of_max_square(row + 1 ,col + 1 ) __UpperCamelCase : List[Any] =update_area_of_max_square(row + 1 ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : Dict =max(largest_square_area[0] ,a_ ) return sub_problem_sol else: return 0 __UpperCamelCase : Union[str, Any] =[0] update_area_of_max_square(0 ,0 ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: def update_area_of_max_square_using_dp_array( a_ ,a_ ,a_ ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __UpperCamelCase : Tuple =update_area_of_max_square_using_dp_array(a_ ,col + 1 ,a_ ) __UpperCamelCase : Optional[int] =update_area_of_max_square_using_dp_array(row + 1 ,col + 1 ,a_ ) __UpperCamelCase : Any =update_area_of_max_square_using_dp_array(row + 1 ,a_ ,a_ ) if mat[row][col]: __UpperCamelCase : Optional[Any] =1 + min([right, diagonal, down] ) __UpperCamelCase : str =max(largest_square_area[0] ,a_ ) __UpperCamelCase : Any =sub_problem_sol return sub_problem_sol else: return 0 __UpperCamelCase : Tuple =[0] __UpperCamelCase : List[Any] =[[-1] * cols for _ in range(a_ )] update_area_of_max_square_using_dp_array(0 ,0 ,a_ ) return largest_square_area[0] def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Dict =[[0] * (cols + 1) for _ in range(rows + 1 )] __UpperCamelCase : int =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Optional[Any] =dp_array[row][col + 1] __UpperCamelCase : int =dp_array[row + 1][col + 1] __UpperCamelCase : Tuple =dp_array[row + 1][col] if mat[row][col] == 1: __UpperCamelCase : Tuple =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Any =max(dp_array[row][col] ,a_ ) else: __UpperCamelCase : Dict =0 return largest_square_area def A ( a_ ,a_ ,a_ ) -> int: __UpperCamelCase : Any =[0] * (cols + 1) __UpperCamelCase : List[Any] =[0] * (cols + 1) __UpperCamelCase : Tuple =0 for row in range(rows - 1 ,-1 ,-1 ): for col in range(cols - 1 ,-1 ,-1 ): __UpperCamelCase : Any =current_row[col + 1] __UpperCamelCase : Optional[Any] =next_row[col + 1] __UpperCamelCase : Union[str, Any] =next_row[col] if mat[row][col] == 1: __UpperCamelCase : Any =1 + min(a_ ,a_ ,a_ ) __UpperCamelCase : Optional[int] =max(current_row[col] ,a_ ) else: __UpperCamelCase : List[str] =0 __UpperCamelCase : Optional[Any] =current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
71
1
from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ :int = logging.get_logger(__name__) A_ :Any = { '''google/efficientnet-b7''': '''https://huggingface.co/google/efficientnet-b7/resolve/main/config.json''', } class __A ( a ): """simple docstring""" UpperCamelCase__ : Optional[Any] ="""efficientnet""" def __init__( self , lowerCamelCase__ = 3 , lowerCamelCase__ = 600 , lowerCamelCase__ = 2.0 , lowerCamelCase__ = 3.1 , lowerCamelCase__ = 8 , lowerCamelCase__ = [3, 3, 5, 3, 5, 5, 3] , lowerCamelCase__ = [32, 16, 24, 40, 80, 112, 192] , lowerCamelCase__ = [16, 24, 40, 80, 112, 192, 320] , lowerCamelCase__ = [] , lowerCamelCase__ = [1, 2, 2, 2, 1, 2, 1] , lowerCamelCase__ = [1, 2, 2, 3, 3, 4, 1] , lowerCamelCase__ = [1, 6, 6, 6, 6, 6, 6] , lowerCamelCase__ = 0.25 , lowerCamelCase__ = "swish" , lowerCamelCase__ = 2560 , lowerCamelCase__ = "mean" , lowerCamelCase__ = 0.02 , lowerCamelCase__ = 0.001 , lowerCamelCase__ = 0.99 , lowerCamelCase__ = 0.5 , lowerCamelCase__ = 0.2 , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : str =num_channels __UpperCamelCase : Any =image_size __UpperCamelCase : List[Any] =width_coefficient __UpperCamelCase : List[Any] =depth_coefficient __UpperCamelCase : Dict =depth_divisor __UpperCamelCase : Tuple =kernel_sizes __UpperCamelCase : Optional[int] =in_channels __UpperCamelCase : Dict =out_channels __UpperCamelCase : Union[str, Any] =depthwise_padding __UpperCamelCase : Optional[Any] =strides __UpperCamelCase : List[str] =num_block_repeats __UpperCamelCase : Any =expand_ratios __UpperCamelCase : Union[str, Any] =squeeze_expansion_ratio __UpperCamelCase : Optional[Any] =hidden_act __UpperCamelCase : int =hidden_dim __UpperCamelCase : Dict =pooling_type __UpperCamelCase : List[Any] =initializer_range __UpperCamelCase : Optional[int] =batch_norm_eps __UpperCamelCase : Dict =batch_norm_momentum __UpperCamelCase : Union[str, Any] =dropout_rate __UpperCamelCase : Dict =drop_connect_rate __UpperCamelCase : List[str] =sum(lowerCamelCase__ ) * 4 class __A ( a ): """simple docstring""" UpperCamelCase__ : Any =version.parse("""1.11""" ) @property def __lowercase ( self ): """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def __lowercase ( self ): """simple docstring""" return 1E-5
71
def A ( a_ ) -> int: __UpperCamelCase : Any =len(a_ ) while cur > 1: # Find the maximum number in arr __UpperCamelCase : Any =arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi __UpperCamelCase : Any =arr[mi::-1] + arr[mi + 1 : len(a_ )] # Reverse whole list __UpperCamelCase : str =arr[cur - 1 :: -1] + arr[cur : len(a_ )] cur -= 1 return arr if __name__ == "__main__": A_ :Dict = input('''Enter numbers separated by a comma:\n''').strip() A_ :Any = [int(item) for item in user_input.split(''',''')] print(pancake_sort(unsorted))
71
1